var/home/core/zuul-output/0000755000175000017500000000000015071214216014524 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015071232721015471 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005414600515071232712017702 0ustar rootrootOct 07 13:45:16 crc systemd[1]: Starting Kubernetes Kubelet... Oct 07 13:45:16 crc restorecon[4724]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 13:45:16 crc restorecon[4724]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 13:45:17 crc restorecon[4724]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 07 13:45:17 crc restorecon[4724]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 07 13:45:18 crc kubenswrapper[4959]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 07 13:45:18 crc kubenswrapper[4959]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 07 13:45:18 crc kubenswrapper[4959]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 07 13:45:18 crc kubenswrapper[4959]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 07 13:45:18 crc kubenswrapper[4959]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 07 13:45:18 crc kubenswrapper[4959]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.372424 4959 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.380977 4959 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381035 4959 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381048 4959 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381059 4959 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381075 4959 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381087 4959 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381139 4959 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381151 4959 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381161 4959 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381171 4959 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381182 4959 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381192 4959 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381202 4959 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381212 4959 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381222 4959 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381232 4959 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381242 4959 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381252 4959 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381262 4959 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381289 4959 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381300 4959 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381310 4959 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381320 4959 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381341 4959 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381353 4959 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381368 4959 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381384 4959 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381395 4959 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381407 4959 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381420 4959 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381430 4959 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381442 4959 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381453 4959 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381464 4959 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381475 4959 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381488 4959 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381501 4959 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381511 4959 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381521 4959 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381532 4959 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381546 4959 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381557 4959 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381567 4959 feature_gate.go:330] unrecognized feature gate: Example Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381577 4959 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381588 4959 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381599 4959 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381609 4959 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381619 4959 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381630 4959 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381640 4959 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381651 4959 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381662 4959 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381674 4959 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381684 4959 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381695 4959 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381705 4959 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381716 4959 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381729 4959 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381740 4959 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381751 4959 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381761 4959 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381773 4959 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381783 4959 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381794 4959 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381804 4959 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381816 4959 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381834 4959 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381847 4959 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381858 4959 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381869 4959 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.381881 4959 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.383686 4959 flags.go:64] FLAG: --address="0.0.0.0" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.383725 4959 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.383747 4959 flags.go:64] FLAG: --anonymous-auth="true" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.383764 4959 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.383781 4959 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.383794 4959 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.383813 4959 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.383830 4959 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.383842 4959 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.383854 4959 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.383868 4959 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.383886 4959 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.383901 4959 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.383913 4959 flags.go:64] FLAG: --cgroup-root="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.383925 4959 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.383937 4959 flags.go:64] FLAG: --client-ca-file="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.383948 4959 flags.go:64] FLAG: --cloud-config="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.383960 4959 flags.go:64] FLAG: --cloud-provider="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.383972 4959 flags.go:64] FLAG: --cluster-dns="[]" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.383993 4959 flags.go:64] FLAG: --cluster-domain="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384005 4959 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384018 4959 flags.go:64] FLAG: --config-dir="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384031 4959 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384045 4959 flags.go:64] FLAG: --container-log-max-files="5" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384063 4959 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384078 4959 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384091 4959 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384144 4959 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384158 4959 flags.go:64] FLAG: --contention-profiling="false" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384171 4959 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384184 4959 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384197 4959 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384209 4959 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384227 4959 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384240 4959 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384252 4959 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384265 4959 flags.go:64] FLAG: --enable-load-reader="false" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384277 4959 flags.go:64] FLAG: --enable-server="true" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384290 4959 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384308 4959 flags.go:64] FLAG: --event-burst="100" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384323 4959 flags.go:64] FLAG: --event-qps="50" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384336 4959 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384348 4959 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384361 4959 flags.go:64] FLAG: --eviction-hard="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384377 4959 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384390 4959 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384401 4959 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384419 4959 flags.go:64] FLAG: --eviction-soft="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384432 4959 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384444 4959 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384457 4959 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384469 4959 flags.go:64] FLAG: --experimental-mounter-path="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384480 4959 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384492 4959 flags.go:64] FLAG: --fail-swap-on="true" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384505 4959 flags.go:64] FLAG: --feature-gates="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384521 4959 flags.go:64] FLAG: --file-check-frequency="20s" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384535 4959 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384549 4959 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384563 4959 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384577 4959 flags.go:64] FLAG: --healthz-port="10248" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384588 4959 flags.go:64] FLAG: --help="false" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384601 4959 flags.go:64] FLAG: --hostname-override="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384613 4959 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384626 4959 flags.go:64] FLAG: --http-check-frequency="20s" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384639 4959 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384651 4959 flags.go:64] FLAG: --image-credential-provider-config="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384663 4959 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384675 4959 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384687 4959 flags.go:64] FLAG: --image-service-endpoint="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384700 4959 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384711 4959 flags.go:64] FLAG: --kube-api-burst="100" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384724 4959 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384738 4959 flags.go:64] FLAG: --kube-api-qps="50" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384750 4959 flags.go:64] FLAG: --kube-reserved="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384763 4959 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384774 4959 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384787 4959 flags.go:64] FLAG: --kubelet-cgroups="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384799 4959 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384810 4959 flags.go:64] FLAG: --lock-file="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384822 4959 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384835 4959 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384847 4959 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384882 4959 flags.go:64] FLAG: --log-json-split-stream="false" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384899 4959 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384912 4959 flags.go:64] FLAG: --log-text-split-stream="false" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384924 4959 flags.go:64] FLAG: --logging-format="text" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384936 4959 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384950 4959 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384961 4959 flags.go:64] FLAG: --manifest-url="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384976 4959 flags.go:64] FLAG: --manifest-url-header="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.384993 4959 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385006 4959 flags.go:64] FLAG: --max-open-files="1000000" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385022 4959 flags.go:64] FLAG: --max-pods="110" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385034 4959 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385046 4959 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385059 4959 flags.go:64] FLAG: --memory-manager-policy="None" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385071 4959 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385083 4959 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385095 4959 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385142 4959 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385176 4959 flags.go:64] FLAG: --node-status-max-images="50" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385189 4959 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385201 4959 flags.go:64] FLAG: --oom-score-adj="-999" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385214 4959 flags.go:64] FLAG: --pod-cidr="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385227 4959 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385246 4959 flags.go:64] FLAG: --pod-manifest-path="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385258 4959 flags.go:64] FLAG: --pod-max-pids="-1" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385271 4959 flags.go:64] FLAG: --pods-per-core="0" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385283 4959 flags.go:64] FLAG: --port="10250" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385295 4959 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385308 4959 flags.go:64] FLAG: --provider-id="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385320 4959 flags.go:64] FLAG: --qos-reserved="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385332 4959 flags.go:64] FLAG: --read-only-port="10255" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385345 4959 flags.go:64] FLAG: --register-node="true" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385357 4959 flags.go:64] FLAG: --register-schedulable="true" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385369 4959 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385391 4959 flags.go:64] FLAG: --registry-burst="10" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385403 4959 flags.go:64] FLAG: --registry-qps="5" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385415 4959 flags.go:64] FLAG: --reserved-cpus="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385429 4959 flags.go:64] FLAG: --reserved-memory="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385446 4959 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385459 4959 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385472 4959 flags.go:64] FLAG: --rotate-certificates="false" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385484 4959 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385496 4959 flags.go:64] FLAG: --runonce="false" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385508 4959 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385522 4959 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385536 4959 flags.go:64] FLAG: --seccomp-default="false" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385547 4959 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385560 4959 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385573 4959 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385586 4959 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385599 4959 flags.go:64] FLAG: --storage-driver-password="root" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385611 4959 flags.go:64] FLAG: --storage-driver-secure="false" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385623 4959 flags.go:64] FLAG: --storage-driver-table="stats" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385637 4959 flags.go:64] FLAG: --storage-driver-user="root" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385650 4959 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385664 4959 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385677 4959 flags.go:64] FLAG: --system-cgroups="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385689 4959 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385709 4959 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385721 4959 flags.go:64] FLAG: --tls-cert-file="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385733 4959 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385751 4959 flags.go:64] FLAG: --tls-min-version="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385764 4959 flags.go:64] FLAG: --tls-private-key-file="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385775 4959 flags.go:64] FLAG: --topology-manager-policy="none" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385787 4959 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385800 4959 flags.go:64] FLAG: --topology-manager-scope="container" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385812 4959 flags.go:64] FLAG: --v="2" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385830 4959 flags.go:64] FLAG: --version="false" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385847 4959 flags.go:64] FLAG: --vmodule="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385862 4959 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.385876 4959 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386242 4959 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386261 4959 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386279 4959 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386292 4959 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386304 4959 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386315 4959 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386326 4959 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386337 4959 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386348 4959 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386359 4959 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386371 4959 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386382 4959 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386393 4959 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386403 4959 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386413 4959 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386426 4959 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386436 4959 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386447 4959 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386458 4959 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386467 4959 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386478 4959 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386492 4959 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386506 4959 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386520 4959 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386535 4959 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386548 4959 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386561 4959 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386572 4959 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386585 4959 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386599 4959 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386610 4959 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386622 4959 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386634 4959 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386644 4959 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386656 4959 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386667 4959 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386678 4959 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386688 4959 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386701 4959 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386712 4959 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386722 4959 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386733 4959 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386745 4959 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386755 4959 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386765 4959 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386776 4959 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386788 4959 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386798 4959 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386808 4959 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386818 4959 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386829 4959 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386841 4959 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386851 4959 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386862 4959 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386873 4959 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386932 4959 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386943 4959 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386955 4959 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386966 4959 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386977 4959 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386987 4959 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.386998 4959 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.387009 4959 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.387019 4959 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.387031 4959 feature_gate.go:330] unrecognized feature gate: Example Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.387041 4959 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.387051 4959 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.387062 4959 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.387073 4959 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.387083 4959 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.387094 4959 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.388080 4959 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.405177 4959 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.405247 4959 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405403 4959 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405427 4959 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405437 4959 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405478 4959 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405490 4959 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405498 4959 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405507 4959 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405516 4959 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405525 4959 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405533 4959 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405542 4959 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405550 4959 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405561 4959 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405575 4959 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405587 4959 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405598 4959 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405608 4959 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405620 4959 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405630 4959 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405639 4959 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405648 4959 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405659 4959 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405671 4959 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405680 4959 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405688 4959 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405697 4959 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405705 4959 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405713 4959 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405721 4959 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405730 4959 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405738 4959 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405747 4959 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405757 4959 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405768 4959 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405782 4959 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405793 4959 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405808 4959 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405820 4959 feature_gate.go:330] unrecognized feature gate: Example Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405830 4959 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405842 4959 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405852 4959 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405861 4959 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405870 4959 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405879 4959 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405887 4959 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405895 4959 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405904 4959 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405913 4959 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405921 4959 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405929 4959 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405937 4959 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405946 4959 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405955 4959 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405963 4959 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405971 4959 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405979 4959 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405987 4959 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.405996 4959 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406004 4959 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406012 4959 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406019 4959 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406027 4959 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406036 4959 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406044 4959 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406052 4959 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406061 4959 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406069 4959 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406077 4959 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406085 4959 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406093 4959 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406139 4959 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.406158 4959 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406432 4959 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406449 4959 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406459 4959 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406468 4959 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406476 4959 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406485 4959 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406494 4959 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406503 4959 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406512 4959 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406521 4959 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406530 4959 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406539 4959 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406547 4959 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406555 4959 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406563 4959 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406572 4959 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406581 4959 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406589 4959 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406597 4959 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406606 4959 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406614 4959 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406622 4959 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406630 4959 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406639 4959 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406647 4959 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406655 4959 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406663 4959 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406671 4959 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406678 4959 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406686 4959 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406697 4959 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406707 4959 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406718 4959 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406727 4959 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406738 4959 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406748 4959 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406756 4959 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406766 4959 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406774 4959 feature_gate.go:330] unrecognized feature gate: Example Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406783 4959 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406791 4959 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406800 4959 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406845 4959 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406854 4959 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406864 4959 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406873 4959 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406881 4959 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406923 4959 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406933 4959 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406941 4959 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406948 4959 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406957 4959 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406965 4959 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406973 4959 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406982 4959 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406990 4959 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.406997 4959 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.407005 4959 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.407013 4959 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.407020 4959 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.407028 4959 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.407035 4959 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.407044 4959 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.407052 4959 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.407062 4959 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.407072 4959 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.407081 4959 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.407092 4959 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.407128 4959 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.407141 4959 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.407157 4959 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.407175 4959 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.407473 4959 server.go:940] "Client rotation is on, will bootstrap in background" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.413575 4959 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.413739 4959 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.415757 4959 server.go:997] "Starting client certificate rotation" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.415800 4959 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.416224 4959 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-15 11:27:42.876923679 +0000 UTC Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.416351 4959 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 2397h42m24.460579652s for next certificate rotation Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.449697 4959 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.452095 4959 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.472650 4959 log.go:25] "Validated CRI v1 runtime API" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.506671 4959 log.go:25] "Validated CRI v1 image API" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.509503 4959 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.515884 4959 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-07-13-40-18-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.515940 4959 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.553262 4959 manager.go:217] Machine: {Timestamp:2025-10-07 13:45:18.547039399 +0000 UTC m=+0.630443804 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:f2d8c8c7-39b4-4371-b614-9cd97280b297 BootID:b4282f39-e355-4eb9-8db9-50fa069893da Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:16:61:39 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:16:61:39 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:df:a8:bc Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:e7:40:28 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:80:de:74 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:f1:87:49 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:6f:92:7e Speed:-1 Mtu:1496} {Name:eth10 MacAddress:36:c5:fc:64:5a:4c Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:6e:ad:25:8f:1e:81 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.553778 4959 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.554032 4959 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.556669 4959 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.557042 4959 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.557156 4959 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.557523 4959 topology_manager.go:138] "Creating topology manager with none policy" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.557544 4959 container_manager_linux.go:303] "Creating device plugin manager" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.558051 4959 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.558147 4959 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.558507 4959 state_mem.go:36] "Initialized new in-memory state store" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.558655 4959 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.562920 4959 kubelet.go:418] "Attempting to sync node with API server" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.562967 4959 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.563089 4959 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.563140 4959 kubelet.go:324] "Adding apiserver pod source" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.563166 4959 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.567290 4959 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.568395 4959 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.569535 4959 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.222:6443: connect: connection refused Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.569573 4959 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.222:6443: connect: connection refused Oct 07 13:45:18 crc kubenswrapper[4959]: E1007 13:45:18.569649 4959 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.222:6443: connect: connection refused" logger="UnhandledError" Oct 07 13:45:18 crc kubenswrapper[4959]: E1007 13:45:18.569774 4959 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.222:6443: connect: connection refused" logger="UnhandledError" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.571161 4959 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.572997 4959 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.573032 4959 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.573043 4959 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.573053 4959 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.573070 4959 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.573081 4959 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.573147 4959 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.573167 4959 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.573179 4959 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.573190 4959 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.573230 4959 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.573244 4959 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.574211 4959 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.574953 4959 server.go:1280] "Started kubelet" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.575923 4959 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.575942 4959 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.575958 4959 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.222:6443: connect: connection refused Oct 07 13:45:18 crc systemd[1]: Started Kubernetes Kubelet. Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.577406 4959 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.581204 4959 server.go:460] "Adding debug handlers to kubelet server" Oct 07 13:45:18 crc kubenswrapper[4959]: E1007 13:45:18.580726 4959 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.222:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186c39728e812190 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-07 13:45:18.574903696 +0000 UTC m=+0.658308031,LastTimestamp:2025-10-07 13:45:18.574903696 +0000 UTC m=+0.658308031,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.582015 4959 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.584374 4959 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.584462 4959 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 18:19:18.638821701 +0000 UTC Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.584847 4959 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 2308h34m0.05398783s for next certificate rotation Oct 07 13:45:18 crc kubenswrapper[4959]: E1007 13:45:18.584760 4959 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.584878 4959 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.584918 4959 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.585482 4959 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 07 13:45:18 crc kubenswrapper[4959]: E1007 13:45:18.586217 4959 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.222:6443: connect: connection refused" interval="200ms" Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.587806 4959 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.222:6443: connect: connection refused Oct 07 13:45:18 crc kubenswrapper[4959]: E1007 13:45:18.587981 4959 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.222:6443: connect: connection refused" logger="UnhandledError" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.591605 4959 factory.go:55] Registering systemd factory Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.591656 4959 factory.go:221] Registration of the systemd container factory successfully Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.592309 4959 factory.go:153] Registering CRI-O factory Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.592369 4959 factory.go:221] Registration of the crio container factory successfully Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.592482 4959 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.592514 4959 factory.go:103] Registering Raw factory Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.592539 4959 manager.go:1196] Started watching for new ooms in manager Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.593433 4959 manager.go:319] Starting recovery of all containers Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.604773 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.604844 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.604871 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.604890 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.604907 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.604924 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.604944 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.604961 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.604981 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.604997 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.605013 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.605032 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.605049 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.605072 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.609603 4959 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.609667 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.609693 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.609742 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.609759 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.609772 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.609787 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.609824 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.609838 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.609852 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.609872 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.609919 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.609939 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.609964 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610051 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610095 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610161 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610177 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610228 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610284 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610328 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610346 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610364 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610408 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610427 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610443 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610494 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610513 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610533 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610578 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610600 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610623 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610672 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610690 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610711 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610752 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610770 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610786 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610827 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610852 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610872 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610918 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610938 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.610956 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611028 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611050 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611093 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611129 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611147 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611192 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611218 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611236 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611253 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611280 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611299 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611350 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611417 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611441 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611470 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611489 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611509 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611565 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611588 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611609 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611637 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611667 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611691 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611755 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611775 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611793 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611821 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611840 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611857 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611874 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611897 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611960 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.611986 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612014 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612033 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612059 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612087 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612126 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612151 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612176 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612197 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612224 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612244 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612263 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612285 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612313 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612335 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612365 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612398 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612420 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612443 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612464 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612492 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612513 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612539 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612560 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612583 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612600 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612617 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612631 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612672 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612686 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612702 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612722 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612740 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612767 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612788 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612806 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612824 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612842 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612863 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612884 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612905 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612924 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612941 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612961 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612981 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.612999 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613024 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613044 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613063 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613081 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613124 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613147 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613174 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613195 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613218 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613240 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613259 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613278 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613296 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613322 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613349 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613373 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613396 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613416 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613435 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613453 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613470 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613487 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613508 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613526 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613546 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613565 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613582 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613600 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613617 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613636 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613655 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613675 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613696 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613715 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613733 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613752 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613775 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613794 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613814 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613834 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613856 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613876 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613896 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613915 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613935 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613955 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.613976 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.614028 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.614051 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.614074 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.614094 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.614133 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.614154 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.614173 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.614191 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.614209 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.614231 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.614252 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.614273 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.614292 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.614311 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.614333 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.614351 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.614371 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.614403 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.614423 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.614442 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.615174 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.615283 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.615327 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.615346 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.615366 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.615381 4959 reconstruct.go:97] "Volume reconstruction finished" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.615393 4959 reconciler.go:26] "Reconciler: start to sync state" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.619483 4959 manager.go:324] Recovery completed Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.634500 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.636219 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.636276 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.636294 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.637529 4959 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.637556 4959 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.637583 4959 state_mem.go:36] "Initialized new in-memory state store" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.649378 4959 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.651310 4959 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.652078 4959 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.652237 4959 kubelet.go:2335] "Starting kubelet main sync loop" Oct 07 13:45:18 crc kubenswrapper[4959]: E1007 13:45:18.652309 4959 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 07 13:45:18 crc kubenswrapper[4959]: W1007 13:45:18.652914 4959 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.222:6443: connect: connection refused Oct 07 13:45:18 crc kubenswrapper[4959]: E1007 13:45:18.653063 4959 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.222:6443: connect: connection refused" logger="UnhandledError" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.656512 4959 policy_none.go:49] "None policy: Start" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.657827 4959 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.657857 4959 state_mem.go:35] "Initializing new in-memory state store" Oct 07 13:45:18 crc kubenswrapper[4959]: E1007 13:45:18.686293 4959 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.721493 4959 manager.go:334] "Starting Device Plugin manager" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.721552 4959 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.721568 4959 server.go:79] "Starting device plugin registration server" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.722122 4959 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.722147 4959 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.722607 4959 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.722698 4959 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.722708 4959 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 07 13:45:18 crc kubenswrapper[4959]: E1007 13:45:18.730373 4959 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.752780 4959 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.752901 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.754587 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.754650 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.754665 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.754869 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.755816 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.755860 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.756245 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.756320 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.756340 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.756604 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.756634 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.756684 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.756695 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.756718 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.756765 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.758194 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.758201 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.758245 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.758486 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.758457 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.758561 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.758690 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.758729 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.758775 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.760049 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.760126 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.760147 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.760058 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.760264 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.760275 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.760371 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.760480 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.760514 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.761671 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.761700 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.761714 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.761733 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.761762 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.761778 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.761887 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.761918 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.763696 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.763728 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.763739 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:18 crc kubenswrapper[4959]: E1007 13:45:18.786980 4959 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.222:6443: connect: connection refused" interval="400ms" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.817871 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.817922 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.817956 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.817983 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.818051 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.818092 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.818190 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.818252 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.818284 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.818314 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.818352 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.818378 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.818406 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.818434 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.818483 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.823002 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.826399 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.826473 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.826496 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.826544 4959 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 07 13:45:18 crc kubenswrapper[4959]: E1007 13:45:18.827314 4959 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.222:6443: connect: connection refused" node="crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.919718 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.919771 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.919795 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.919813 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.919834 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.919853 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.919871 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.919887 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.919906 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.919926 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.919947 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.919965 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.919987 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.920005 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.920027 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.920150 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.920358 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.920384 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.920429 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.920428 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.920461 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.920512 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.920524 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.920541 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.920577 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.920578 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.920611 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.920638 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.920663 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 07 13:45:18 crc kubenswrapper[4959]: I1007 13:45:18.920666 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 13:45:19 crc kubenswrapper[4959]: I1007 13:45:19.029165 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:19 crc kubenswrapper[4959]: I1007 13:45:19.030875 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:19 crc kubenswrapper[4959]: I1007 13:45:19.030946 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:19 crc kubenswrapper[4959]: I1007 13:45:19.030969 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:19 crc kubenswrapper[4959]: I1007 13:45:19.031014 4959 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 07 13:45:19 crc kubenswrapper[4959]: E1007 13:45:19.031854 4959 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.222:6443: connect: connection refused" node="crc" Oct 07 13:45:19 crc kubenswrapper[4959]: I1007 13:45:19.081215 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 13:45:19 crc kubenswrapper[4959]: I1007 13:45:19.087842 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 07 13:45:19 crc kubenswrapper[4959]: I1007 13:45:19.114675 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 07 13:45:19 crc kubenswrapper[4959]: I1007 13:45:19.121914 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 13:45:19 crc kubenswrapper[4959]: I1007 13:45:19.127214 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 13:45:19 crc kubenswrapper[4959]: W1007 13:45:19.138709 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-4811576a4dfb67883af65a3286c8acf756d8995969c01b1600f70e628642d95d WatchSource:0}: Error finding container 4811576a4dfb67883af65a3286c8acf756d8995969c01b1600f70e628642d95d: Status 404 returned error can't find the container with id 4811576a4dfb67883af65a3286c8acf756d8995969c01b1600f70e628642d95d Oct 07 13:45:19 crc kubenswrapper[4959]: W1007 13:45:19.154614 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-90d435f2069454704fdf30bfc794217a84e19e857cc10304c47ca1bc1029a0d3 WatchSource:0}: Error finding container 90d435f2069454704fdf30bfc794217a84e19e857cc10304c47ca1bc1029a0d3: Status 404 returned error can't find the container with id 90d435f2069454704fdf30bfc794217a84e19e857cc10304c47ca1bc1029a0d3 Oct 07 13:45:19 crc kubenswrapper[4959]: W1007 13:45:19.160127 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-20dd71222bf9e937c387beb900467bbd49134230a5ca59ec5bbe4cdaf3a1c9ac WatchSource:0}: Error finding container 20dd71222bf9e937c387beb900467bbd49134230a5ca59ec5bbe4cdaf3a1c9ac: Status 404 returned error can't find the container with id 20dd71222bf9e937c387beb900467bbd49134230a5ca59ec5bbe4cdaf3a1c9ac Oct 07 13:45:19 crc kubenswrapper[4959]: W1007 13:45:19.170651 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-0757b09b5002af001175530e424142700b2b537dc346c8da34be71838f5f2ff7 WatchSource:0}: Error finding container 0757b09b5002af001175530e424142700b2b537dc346c8da34be71838f5f2ff7: Status 404 returned error can't find the container with id 0757b09b5002af001175530e424142700b2b537dc346c8da34be71838f5f2ff7 Oct 07 13:45:19 crc kubenswrapper[4959]: E1007 13:45:19.188702 4959 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.222:6443: connect: connection refused" interval="800ms" Oct 07 13:45:19 crc kubenswrapper[4959]: W1007 13:45:19.403025 4959 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.222:6443: connect: connection refused Oct 07 13:45:19 crc kubenswrapper[4959]: E1007 13:45:19.403127 4959 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.222:6443: connect: connection refused" logger="UnhandledError" Oct 07 13:45:19 crc kubenswrapper[4959]: W1007 13:45:19.425872 4959 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.222:6443: connect: connection refused Oct 07 13:45:19 crc kubenswrapper[4959]: E1007 13:45:19.425931 4959 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.222:6443: connect: connection refused" logger="UnhandledError" Oct 07 13:45:19 crc kubenswrapper[4959]: I1007 13:45:19.432313 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:19 crc kubenswrapper[4959]: I1007 13:45:19.433302 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:19 crc kubenswrapper[4959]: I1007 13:45:19.433334 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:19 crc kubenswrapper[4959]: I1007 13:45:19.433342 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:19 crc kubenswrapper[4959]: I1007 13:45:19.433366 4959 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 07 13:45:19 crc kubenswrapper[4959]: E1007 13:45:19.433906 4959 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.222:6443: connect: connection refused" node="crc" Oct 07 13:45:19 crc kubenswrapper[4959]: I1007 13:45:19.577938 4959 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.222:6443: connect: connection refused Oct 07 13:45:19 crc kubenswrapper[4959]: I1007 13:45:19.660368 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3392861d8e9c95ecf826dadfa0079c6085d008e64f819b84180f8959c22739fc"} Oct 07 13:45:19 crc kubenswrapper[4959]: I1007 13:45:19.661374 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0757b09b5002af001175530e424142700b2b537dc346c8da34be71838f5f2ff7"} Oct 07 13:45:19 crc kubenswrapper[4959]: I1007 13:45:19.662447 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"20dd71222bf9e937c387beb900467bbd49134230a5ca59ec5bbe4cdaf3a1c9ac"} Oct 07 13:45:19 crc kubenswrapper[4959]: I1007 13:45:19.663709 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"90d435f2069454704fdf30bfc794217a84e19e857cc10304c47ca1bc1029a0d3"} Oct 07 13:45:19 crc kubenswrapper[4959]: I1007 13:45:19.664902 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"4811576a4dfb67883af65a3286c8acf756d8995969c01b1600f70e628642d95d"} Oct 07 13:45:19 crc kubenswrapper[4959]: W1007 13:45:19.712273 4959 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.222:6443: connect: connection refused Oct 07 13:45:19 crc kubenswrapper[4959]: E1007 13:45:19.712490 4959 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.222:6443: connect: connection refused" logger="UnhandledError" Oct 07 13:45:19 crc kubenswrapper[4959]: W1007 13:45:19.764507 4959 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.222:6443: connect: connection refused Oct 07 13:45:19 crc kubenswrapper[4959]: E1007 13:45:19.764654 4959 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.222:6443: connect: connection refused" logger="UnhandledError" Oct 07 13:45:19 crc kubenswrapper[4959]: E1007 13:45:19.989710 4959 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.222:6443: connect: connection refused" interval="1.6s" Oct 07 13:45:20 crc kubenswrapper[4959]: E1007 13:45:20.006534 4959 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.222:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186c39728e812190 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-07 13:45:18.574903696 +0000 UTC m=+0.658308031,LastTimestamp:2025-10-07 13:45:18.574903696 +0000 UTC m=+0.658308031,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.234040 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.235684 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.235725 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.235739 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.235769 4959 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 07 13:45:20 crc kubenswrapper[4959]: E1007 13:45:20.236353 4959 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.222:6443: connect: connection refused" node="crc" Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.577561 4959 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.222:6443: connect: connection refused Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.671266 4959 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a" exitCode=0 Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.671381 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a"} Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.671504 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.674045 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.674181 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.674209 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.675422 4959 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c" exitCode=0 Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.675543 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c"} Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.675571 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.677255 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.677315 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.677336 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.681152 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.681233 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7"} Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.681300 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6"} Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.681334 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40"} Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.681351 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.681361 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038"} Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.682907 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.682911 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.682999 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.683018 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.682960 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.683068 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.685134 4959 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb" exitCode=0 Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.685238 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb"} Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.685298 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.686597 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.686677 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.686704 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.687578 4959 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="b6d82be64b945bd47d6149c695f43152998644d31f0103bfff861199d5efb191" exitCode=0 Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.687624 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"b6d82be64b945bd47d6149c695f43152998644d31f0103bfff861199d5efb191"} Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.687696 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.689055 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.689091 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:20 crc kubenswrapper[4959]: I1007 13:45:20.689129 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:21 crc kubenswrapper[4959]: W1007 13:45:21.204438 4959 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.222:6443: connect: connection refused Oct 07 13:45:21 crc kubenswrapper[4959]: E1007 13:45:21.204533 4959 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.222:6443: connect: connection refused" logger="UnhandledError" Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.578142 4959 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.222:6443: connect: connection refused Oct 07 13:45:21 crc kubenswrapper[4959]: E1007 13:45:21.590353 4959 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.222:6443: connect: connection refused" interval="3.2s" Oct 07 13:45:21 crc kubenswrapper[4959]: W1007 13:45:21.671554 4959 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.222:6443: connect: connection refused Oct 07 13:45:21 crc kubenswrapper[4959]: E1007 13:45:21.671671 4959 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.222:6443: connect: connection refused" logger="UnhandledError" Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.696923 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9b500df391c3b1ef8f9689a04e58c19e603667507a68fb8d65ea768690425085"} Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.696988 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c6dab7bc77a409dca97cffd302d0086021bf51146e8ec69a53cbddb87a4026aa"} Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.697008 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e3ecb7da009b7ed9e5c0412125be08cf486bf4f0e74480f2843877b0816944fe"} Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.697012 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.698371 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.698429 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.698449 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.702297 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0"} Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.702354 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2"} Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.702374 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9"} Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.702391 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544"} Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.705815 4959 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317" exitCode=0 Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.705893 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317"} Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.706065 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.707271 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.707314 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.707332 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.711519 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.712069 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.712183 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"ae966401c1e804da8649ca9c014cfc2eed8dbb12a0576cf90b6c7a3526513601"} Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.712509 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.712557 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.712573 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.712994 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.713036 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.713048 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:21 crc kubenswrapper[4959]: W1007 13:45:21.817088 4959 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.222:6443: connect: connection refused Oct 07 13:45:21 crc kubenswrapper[4959]: E1007 13:45:21.817241 4959 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.222:6443: connect: connection refused" logger="UnhandledError" Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.836904 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.839223 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.839267 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.839281 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:21 crc kubenswrapper[4959]: I1007 13:45:21.839307 4959 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 07 13:45:21 crc kubenswrapper[4959]: E1007 13:45:21.839881 4959 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.222:6443: connect: connection refused" node="crc" Oct 07 13:45:22 crc kubenswrapper[4959]: I1007 13:45:22.087636 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 13:45:22 crc kubenswrapper[4959]: I1007 13:45:22.529257 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 13:45:22 crc kubenswrapper[4959]: I1007 13:45:22.535884 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 13:45:22 crc kubenswrapper[4959]: I1007 13:45:22.717006 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58"} Oct 07 13:45:22 crc kubenswrapper[4959]: I1007 13:45:22.717073 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:22 crc kubenswrapper[4959]: I1007 13:45:22.717929 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:22 crc kubenswrapper[4959]: I1007 13:45:22.717976 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:22 crc kubenswrapper[4959]: I1007 13:45:22.717996 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:22 crc kubenswrapper[4959]: I1007 13:45:22.719504 4959 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a" exitCode=0 Oct 07 13:45:22 crc kubenswrapper[4959]: I1007 13:45:22.719564 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:22 crc kubenswrapper[4959]: I1007 13:45:22.719638 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:22 crc kubenswrapper[4959]: I1007 13:45:22.719661 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a"} Oct 07 13:45:22 crc kubenswrapper[4959]: I1007 13:45:22.719667 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:22 crc kubenswrapper[4959]: I1007 13:45:22.719820 4959 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 07 13:45:22 crc kubenswrapper[4959]: I1007 13:45:22.719880 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:22 crc kubenswrapper[4959]: I1007 13:45:22.720414 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:22 crc kubenswrapper[4959]: I1007 13:45:22.720470 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:22 crc kubenswrapper[4959]: I1007 13:45:22.720492 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:22 crc kubenswrapper[4959]: I1007 13:45:22.720897 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:22 crc kubenswrapper[4959]: I1007 13:45:22.720927 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:22 crc kubenswrapper[4959]: I1007 13:45:22.720941 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:22 crc kubenswrapper[4959]: I1007 13:45:22.720958 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:22 crc kubenswrapper[4959]: I1007 13:45:22.721001 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:22 crc kubenswrapper[4959]: I1007 13:45:22.721021 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:22 crc kubenswrapper[4959]: I1007 13:45:22.721418 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:22 crc kubenswrapper[4959]: I1007 13:45:22.721431 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:22 crc kubenswrapper[4959]: I1007 13:45:22.721439 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:23 crc kubenswrapper[4959]: I1007 13:45:23.173865 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 13:45:23 crc kubenswrapper[4959]: I1007 13:45:23.716550 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 13:45:23 crc kubenswrapper[4959]: I1007 13:45:23.724887 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd"} Oct 07 13:45:23 crc kubenswrapper[4959]: I1007 13:45:23.724960 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398"} Oct 07 13:45:23 crc kubenswrapper[4959]: I1007 13:45:23.724983 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057"} Oct 07 13:45:23 crc kubenswrapper[4959]: I1007 13:45:23.724992 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:23 crc kubenswrapper[4959]: I1007 13:45:23.725002 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b"} Oct 07 13:45:23 crc kubenswrapper[4959]: I1007 13:45:23.724907 4959 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 07 13:45:23 crc kubenswrapper[4959]: I1007 13:45:23.725084 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:23 crc kubenswrapper[4959]: I1007 13:45:23.725840 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:23 crc kubenswrapper[4959]: I1007 13:45:23.725867 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:23 crc kubenswrapper[4959]: I1007 13:45:23.725877 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:23 crc kubenswrapper[4959]: I1007 13:45:23.726216 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:23 crc kubenswrapper[4959]: I1007 13:45:23.726246 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:23 crc kubenswrapper[4959]: I1007 13:45:23.726257 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:24 crc kubenswrapper[4959]: I1007 13:45:24.732391 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30"} Oct 07 13:45:24 crc kubenswrapper[4959]: I1007 13:45:24.732464 4959 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 07 13:45:24 crc kubenswrapper[4959]: I1007 13:45:24.732478 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:24 crc kubenswrapper[4959]: I1007 13:45:24.732523 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:24 crc kubenswrapper[4959]: I1007 13:45:24.732634 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:24 crc kubenswrapper[4959]: I1007 13:45:24.733576 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:24 crc kubenswrapper[4959]: I1007 13:45:24.733628 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:24 crc kubenswrapper[4959]: I1007 13:45:24.733643 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:24 crc kubenswrapper[4959]: I1007 13:45:24.733951 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:24 crc kubenswrapper[4959]: I1007 13:45:24.734011 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:24 crc kubenswrapper[4959]: I1007 13:45:24.734032 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:24 crc kubenswrapper[4959]: I1007 13:45:24.734050 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:24 crc kubenswrapper[4959]: I1007 13:45:24.734085 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:24 crc kubenswrapper[4959]: I1007 13:45:24.734141 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:25 crc kubenswrapper[4959]: I1007 13:45:25.040344 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:25 crc kubenswrapper[4959]: I1007 13:45:25.042199 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:25 crc kubenswrapper[4959]: I1007 13:45:25.042251 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:25 crc kubenswrapper[4959]: I1007 13:45:25.042268 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:25 crc kubenswrapper[4959]: I1007 13:45:25.042300 4959 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 07 13:45:25 crc kubenswrapper[4959]: I1007 13:45:25.082653 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 13:45:25 crc kubenswrapper[4959]: I1007 13:45:25.244240 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 13:45:25 crc kubenswrapper[4959]: I1007 13:45:25.735180 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:25 crc kubenswrapper[4959]: I1007 13:45:25.735323 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:25 crc kubenswrapper[4959]: I1007 13:45:25.735348 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:25 crc kubenswrapper[4959]: I1007 13:45:25.736560 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:25 crc kubenswrapper[4959]: I1007 13:45:25.736620 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:25 crc kubenswrapper[4959]: I1007 13:45:25.736701 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:25 crc kubenswrapper[4959]: I1007 13:45:25.736858 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:25 crc kubenswrapper[4959]: I1007 13:45:25.736877 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:25 crc kubenswrapper[4959]: I1007 13:45:25.736919 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:25 crc kubenswrapper[4959]: I1007 13:45:25.736937 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:25 crc kubenswrapper[4959]: I1007 13:45:25.736892 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:25 crc kubenswrapper[4959]: I1007 13:45:25.736996 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:25 crc kubenswrapper[4959]: I1007 13:45:25.906437 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 13:45:26 crc kubenswrapper[4959]: I1007 13:45:26.738569 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:26 crc kubenswrapper[4959]: I1007 13:45:26.739869 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:26 crc kubenswrapper[4959]: I1007 13:45:26.739935 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:26 crc kubenswrapper[4959]: I1007 13:45:26.739957 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:28 crc kubenswrapper[4959]: I1007 13:45:28.083346 4959 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 07 13:45:28 crc kubenswrapper[4959]: I1007 13:45:28.083485 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 07 13:45:28 crc kubenswrapper[4959]: I1007 13:45:28.389809 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 13:45:28 crc kubenswrapper[4959]: I1007 13:45:28.390086 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:28 crc kubenswrapper[4959]: I1007 13:45:28.391621 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:28 crc kubenswrapper[4959]: I1007 13:45:28.391685 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:28 crc kubenswrapper[4959]: I1007 13:45:28.391704 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:28 crc kubenswrapper[4959]: E1007 13:45:28.730479 4959 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 07 13:45:29 crc kubenswrapper[4959]: I1007 13:45:29.649616 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 07 13:45:29 crc kubenswrapper[4959]: I1007 13:45:29.649840 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:29 crc kubenswrapper[4959]: I1007 13:45:29.651276 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:29 crc kubenswrapper[4959]: I1007 13:45:29.651342 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:29 crc kubenswrapper[4959]: I1007 13:45:29.651364 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:32 crc kubenswrapper[4959]: I1007 13:45:32.092349 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 13:45:32 crc kubenswrapper[4959]: I1007 13:45:32.092562 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:32 crc kubenswrapper[4959]: I1007 13:45:32.094024 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:32 crc kubenswrapper[4959]: I1007 13:45:32.094090 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:32 crc kubenswrapper[4959]: I1007 13:45:32.094120 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:32 crc kubenswrapper[4959]: W1007 13:45:32.418749 4959 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 07 13:45:32 crc kubenswrapper[4959]: I1007 13:45:32.418834 4959 trace.go:236] Trace[2138418861]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (07-Oct-2025 13:45:22.417) (total time: 10001ms): Oct 07 13:45:32 crc kubenswrapper[4959]: Trace[2138418861]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (13:45:32.418) Oct 07 13:45:32 crc kubenswrapper[4959]: Trace[2138418861]: [10.001208547s] [10.001208547s] END Oct 07 13:45:32 crc kubenswrapper[4959]: E1007 13:45:32.418854 4959 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 07 13:45:32 crc kubenswrapper[4959]: I1007 13:45:32.467717 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 07 13:45:32 crc kubenswrapper[4959]: I1007 13:45:32.467975 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:32 crc kubenswrapper[4959]: I1007 13:45:32.470336 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:32 crc kubenswrapper[4959]: I1007 13:45:32.470407 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:32 crc kubenswrapper[4959]: I1007 13:45:32.470428 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:32 crc kubenswrapper[4959]: I1007 13:45:32.578354 4959 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 07 13:45:32 crc kubenswrapper[4959]: I1007 13:45:32.670179 4959 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 07 13:45:32 crc kubenswrapper[4959]: I1007 13:45:32.670234 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 07 13:45:32 crc kubenswrapper[4959]: I1007 13:45:32.689149 4959 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 07 13:45:32 crc kubenswrapper[4959]: I1007 13:45:32.689205 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 07 13:45:32 crc kubenswrapper[4959]: I1007 13:45:32.755091 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 07 13:45:32 crc kubenswrapper[4959]: I1007 13:45:32.762323 4959 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58" exitCode=255 Oct 07 13:45:32 crc kubenswrapper[4959]: I1007 13:45:32.762385 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58"} Oct 07 13:45:32 crc kubenswrapper[4959]: I1007 13:45:32.762587 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:32 crc kubenswrapper[4959]: I1007 13:45:32.763499 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:32 crc kubenswrapper[4959]: I1007 13:45:32.763537 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:32 crc kubenswrapper[4959]: I1007 13:45:32.763551 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:32 crc kubenswrapper[4959]: I1007 13:45:32.764167 4959 scope.go:117] "RemoveContainer" containerID="cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58" Oct 07 13:45:33 crc kubenswrapper[4959]: I1007 13:45:33.181710 4959 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]log ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]etcd ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/openshift.io-api-request-count-filter ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/openshift.io-startkubeinformers ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/generic-apiserver-start-informers ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/priority-and-fairness-config-consumer ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/priority-and-fairness-filter ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/start-apiextensions-informers ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/start-apiextensions-controllers ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/crd-informer-synced ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/start-system-namespaces-controller ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/start-cluster-authentication-info-controller ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/start-legacy-token-tracking-controller ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/start-service-ip-repair-controllers ok Oct 07 13:45:33 crc kubenswrapper[4959]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Oct 07 13:45:33 crc kubenswrapper[4959]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/priority-and-fairness-config-producer ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/bootstrap-controller ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/start-kube-aggregator-informers ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/apiservice-status-local-available-controller ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/apiservice-status-remote-available-controller ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/apiservice-registration-controller ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/apiservice-wait-for-first-sync ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/apiservice-discovery-controller ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/kube-apiserver-autoregistration ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]autoregister-completion ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/apiservice-openapi-controller ok Oct 07 13:45:33 crc kubenswrapper[4959]: [+]poststarthook/apiservice-openapiv3-controller ok Oct 07 13:45:33 crc kubenswrapper[4959]: livez check failed Oct 07 13:45:33 crc kubenswrapper[4959]: I1007 13:45:33.181820 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 13:45:33 crc kubenswrapper[4959]: I1007 13:45:33.767051 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 07 13:45:33 crc kubenswrapper[4959]: I1007 13:45:33.768671 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80"} Oct 07 13:45:33 crc kubenswrapper[4959]: I1007 13:45:33.768843 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:33 crc kubenswrapper[4959]: I1007 13:45:33.769848 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:33 crc kubenswrapper[4959]: I1007 13:45:33.769884 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:33 crc kubenswrapper[4959]: I1007 13:45:33.769896 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:35 crc kubenswrapper[4959]: I1007 13:45:35.244802 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 13:45:35 crc kubenswrapper[4959]: I1007 13:45:35.245036 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:35 crc kubenswrapper[4959]: I1007 13:45:35.246554 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:35 crc kubenswrapper[4959]: I1007 13:45:35.246608 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:35 crc kubenswrapper[4959]: I1007 13:45:35.246626 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:35 crc kubenswrapper[4959]: I1007 13:45:35.910408 4959 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 07 13:45:37 crc kubenswrapper[4959]: E1007 13:45:37.668409 4959 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 07 13:45:37 crc kubenswrapper[4959]: I1007 13:45:37.670164 4959 trace.go:236] Trace[242007907]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (07-Oct-2025 13:45:27.365) (total time: 10304ms): Oct 07 13:45:37 crc kubenswrapper[4959]: Trace[242007907]: ---"Objects listed" error: 10304ms (13:45:37.670) Oct 07 13:45:37 crc kubenswrapper[4959]: Trace[242007907]: [10.304382027s] [10.304382027s] END Oct 07 13:45:37 crc kubenswrapper[4959]: I1007 13:45:37.670195 4959 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 07 13:45:37 crc kubenswrapper[4959]: I1007 13:45:37.670376 4959 trace.go:236] Trace[1521376125]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (07-Oct-2025 13:45:26.181) (total time: 11488ms): Oct 07 13:45:37 crc kubenswrapper[4959]: Trace[1521376125]: ---"Objects listed" error: 11488ms (13:45:37.670) Oct 07 13:45:37 crc kubenswrapper[4959]: Trace[1521376125]: [11.488903113s] [11.488903113s] END Oct 07 13:45:37 crc kubenswrapper[4959]: I1007 13:45:37.670411 4959 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 07 13:45:37 crc kubenswrapper[4959]: I1007 13:45:37.671089 4959 trace.go:236] Trace[520624934]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (07-Oct-2025 13:45:27.349) (total time: 10321ms): Oct 07 13:45:37 crc kubenswrapper[4959]: Trace[520624934]: ---"Objects listed" error: 10321ms (13:45:37.670) Oct 07 13:45:37 crc kubenswrapper[4959]: Trace[520624934]: [10.321330916s] [10.321330916s] END Oct 07 13:45:37 crc kubenswrapper[4959]: I1007 13:45:37.671134 4959 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 07 13:45:37 crc kubenswrapper[4959]: I1007 13:45:37.674032 4959 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 07 13:45:37 crc kubenswrapper[4959]: E1007 13:45:37.675617 4959 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.084372 4959 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.084447 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.180997 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.185246 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.574367 4959 apiserver.go:52] "Watching apiserver" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.577439 4959 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.577742 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-apiserver/kube-apiserver-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb"] Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.578193 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.578243 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.578211 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.578276 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.578283 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.578605 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 13:45:38 crc kubenswrapper[4959]: E1007 13:45:38.578935 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:45:38 crc kubenswrapper[4959]: E1007 13:45:38.578960 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:45:38 crc kubenswrapper[4959]: E1007 13:45:38.579165 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.580330 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.581261 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.581417 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.582457 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.582455 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.582956 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.583532 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.584398 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.588960 4959 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.590753 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.642634 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.665265 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.680462 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.680540 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.680585 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.680619 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.680651 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.680680 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.680710 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.680740 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.680772 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.680806 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.680837 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.680868 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.680900 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.680937 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.680981 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.680968 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681008 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681068 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681135 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681172 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681220 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681238 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681261 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681296 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681330 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681366 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681402 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681438 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681436 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681474 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681449 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681513 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681586 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681582 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681620 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681655 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681659 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681690 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681726 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681768 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681802 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681839 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681873 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681908 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681948 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.681980 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682012 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682030 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682087 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682147 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682190 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682227 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682264 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682341 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682376 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682446 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682480 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682511 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682546 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682579 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682612 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682645 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682683 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682716 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682751 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682786 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682823 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682867 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682947 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682982 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683019 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683052 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683093 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683152 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683189 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683221 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683258 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683291 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683327 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683362 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683396 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683430 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683463 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683498 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683533 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683566 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683602 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683732 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683770 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683804 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683845 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683879 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683913 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683947 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.684055 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.684094 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.684297 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.684334 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.684369 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.684441 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.684479 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.684514 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.684550 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.684584 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.684619 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.684654 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.684694 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.684727 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.684765 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.684802 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.684841 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.684877 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.684914 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.684948 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.684985 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.685020 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.685056 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.685094 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.685167 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.685206 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.685242 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.685276 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.685315 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.685353 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.685389 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.685426 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.685463 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.685500 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.685562 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.685599 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.685635 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.685673 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.685708 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.685744 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.685779 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.685815 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.685852 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.685888 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.685926 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.685962 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.685996 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686035 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686071 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686137 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686178 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686225 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686262 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686300 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686336 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686372 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686410 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686447 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686482 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686525 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686569 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686606 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686655 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686704 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686742 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686780 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686821 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686858 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686896 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686934 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686971 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687009 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687046 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687084 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687148 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687186 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687226 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687266 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687304 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687341 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687378 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687417 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687460 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687447 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687498 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687540 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687576 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687612 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687649 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687690 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687728 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687766 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687804 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687846 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687882 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687923 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687962 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687997 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.688035 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.688077 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.688265 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.688315 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.688406 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.688445 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.688484 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.688523 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.688560 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.688599 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.688642 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.688682 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.688721 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.689234 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.689291 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.689337 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.689385 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.689438 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.689481 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.689551 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.689599 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.689637 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.689679 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.689722 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.689765 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.689803 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.689846 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.690376 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.690547 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.690579 4959 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.690602 4959 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.690667 4959 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.690689 4959 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.690753 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.690778 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.690842 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682082 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.690824 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682217 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682254 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682292 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682367 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682444 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682585 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682594 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682764 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682880 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682947 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.682987 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683078 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683266 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683440 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683531 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683600 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683812 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.683885 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.684017 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.684033 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.684407 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.684466 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.684755 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.684912 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.685091 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.684988 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.685373 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686172 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686224 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686381 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686680 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686706 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.686824 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687242 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687311 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687649 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687707 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.687885 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.688034 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.688212 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.688539 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.688601 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.688844 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.689211 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.689259 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.689276 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.689315 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.689563 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.690465 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.690500 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.690586 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.690718 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.691057 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.691604 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.692078 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.692155 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.692372 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.692502 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.692614 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.692618 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.692966 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.693063 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.693348 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.693202 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.693490 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.693661 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.693707 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.693875 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.694502 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.694724 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.694828 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.694932 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.694965 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.694974 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.694997 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.695332 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.695438 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.695533 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.695637 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.695679 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.695700 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.695715 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.696031 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.696062 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.696268 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.696425 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: E1007 13:45:38.696535 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:45:39.196513992 +0000 UTC m=+21.279918317 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.696527 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.696625 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.696644 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.696916 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.697048 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.697136 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.697149 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.697283 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.697305 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.697279 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.697721 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.697742 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.697797 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.697740 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.697932 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: E1007 13:45:38.697954 4959 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 13:45:38 crc kubenswrapper[4959]: E1007 13:45:38.698004 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 13:45:39.1979937 +0000 UTC m=+21.281398025 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.698373 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.698393 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.698470 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.698555 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.698589 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.698724 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.698936 4959 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.699010 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.699169 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.699284 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: E1007 13:45:38.699357 4959 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 13:45:38 crc kubenswrapper[4959]: E1007 13:45:38.699392 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 13:45:39.199382255 +0000 UTC m=+21.282786580 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.699547 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.699850 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.700078 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.700325 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.700366 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.700542 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.700599 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.700740 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.700792 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.700859 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.701278 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.701385 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.701931 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.702258 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.702273 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.702384 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.701442 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.702880 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.703204 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.703277 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.703322 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.703374 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.704056 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.704439 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.704754 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.705164 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.705455 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.705580 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.706049 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.706324 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.706618 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.706864 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.706952 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.707570 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.707565 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.707995 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.708076 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.709826 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.710180 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.710729 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.710951 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.711188 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.711318 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.711334 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.711608 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.712019 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.712435 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.712755 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.713482 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.713501 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.713994 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.714644 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.715209 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.715220 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.715553 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.715721 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.716115 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.716293 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.717173 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.717942 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 13:45:38 crc kubenswrapper[4959]: E1007 13:45:38.718216 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 13:45:38 crc kubenswrapper[4959]: E1007 13:45:38.718235 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 13:45:38 crc kubenswrapper[4959]: E1007 13:45:38.718248 4959 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:45:38 crc kubenswrapper[4959]: E1007 13:45:38.718305 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 13:45:39.218288385 +0000 UTC m=+21.301692710 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.719028 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.719914 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.720320 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.723334 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.724293 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.724329 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.724598 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.724814 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.725064 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.726713 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.726782 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.726865 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.727179 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.728422 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: E1007 13:45:38.729456 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 13:45:38 crc kubenswrapper[4959]: E1007 13:45:38.729695 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 13:45:38 crc kubenswrapper[4959]: E1007 13:45:38.729861 4959 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:45:38 crc kubenswrapper[4959]: E1007 13:45:38.730085 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 13:45:39.230051874 +0000 UTC m=+21.313456229 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.732561 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.733413 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.734086 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.738681 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.739242 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.748444 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.754674 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.757944 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.758836 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.761935 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.766764 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.767445 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.775679 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.785534 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 13:45:38 crc kubenswrapper[4959]: E1007 13:45:38.788277 4959 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.791812 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.791861 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.791917 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.791930 4959 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.791941 4959 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.791951 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.791960 4959 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.791970 4959 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.791979 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.791989 4959 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792000 4959 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792008 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792017 4959 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792026 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792036 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792045 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792047 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792054 4959 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792077 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792134 4959 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792154 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792170 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792187 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792201 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792215 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792228 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792242 4959 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792255 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792269 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792281 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792293 4959 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792306 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792320 4959 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792332 4959 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792345 4959 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792357 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792370 4959 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792382 4959 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792394 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792406 4959 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792419 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792435 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792450 4959 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792463 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792476 4959 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792488 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792501 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792513 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792530 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792542 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792555 4959 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792569 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792583 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792595 4959 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792608 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792622 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792635 4959 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792650 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792663 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792675 4959 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792688 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792700 4959 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792712 4959 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792724 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792735 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792746 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792758 4959 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792770 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792782 4959 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792793 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792805 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792818 4959 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792831 4959 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792844 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792857 4959 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792869 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792882 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792893 4959 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792905 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792917 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792929 4959 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792941 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792954 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792971 4959 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792983 4959 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.792996 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793008 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793019 4959 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793031 4959 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793042 4959 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793054 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793068 4959 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793083 4959 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793114 4959 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793128 4959 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793142 4959 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793153 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793165 4959 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793177 4959 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793189 4959 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793200 4959 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793213 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793225 4959 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793237 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793250 4959 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793261 4959 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793273 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793285 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793297 4959 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793308 4959 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793320 4959 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793332 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793345 4959 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793357 4959 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793369 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793383 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793397 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793408 4959 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793421 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793435 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793446 4959 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793457 4959 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793468 4959 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793480 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793492 4959 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793504 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793517 4959 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793529 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793541 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793552 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793565 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793598 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793610 4959 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793622 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793634 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793646 4959 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793658 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793670 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793681 4959 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793694 4959 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793707 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793719 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793730 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793742 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793755 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793766 4959 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793779 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793790 4959 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793804 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793817 4959 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793830 4959 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793842 4959 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793855 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793867 4959 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793878 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793891 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793903 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793915 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793927 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793938 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793950 4959 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793962 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793974 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793985 4959 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.793997 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.794008 4959 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.794021 4959 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.794033 4959 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.794045 4959 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.794057 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.794070 4959 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.794084 4959 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.794109 4959 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.794122 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.794134 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.794147 4959 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.794159 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.794170 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.794181 4959 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.794194 4959 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.794206 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.794218 4959 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.794229 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.794240 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.794252 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.794266 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.794278 4959 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.794290 4959 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.794301 4959 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.800594 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.812387 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.829025 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.846510 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.860210 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.876198 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.894790 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.915001 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 07 13:45:38 crc kubenswrapper[4959]: I1007 13:45:38.924135 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 07 13:45:38 crc kubenswrapper[4959]: W1007 13:45:38.930473 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-c3f72d6aa3541a3b359c575478b733a588d9765284fc2f2db84fd181311cbee5 WatchSource:0}: Error finding container c3f72d6aa3541a3b359c575478b733a588d9765284fc2f2db84fd181311cbee5: Status 404 returned error can't find the container with id c3f72d6aa3541a3b359c575478b733a588d9765284fc2f2db84fd181311cbee5 Oct 07 13:45:38 crc kubenswrapper[4959]: W1007 13:45:38.940541 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-6647be4975a863bd92335c269adc5ea0936cd2ac8296b9add0b76702569a26cc WatchSource:0}: Error finding container 6647be4975a863bd92335c269adc5ea0936cd2ac8296b9add0b76702569a26cc: Status 404 returned error can't find the container with id 6647be4975a863bd92335c269adc5ea0936cd2ac8296b9add0b76702569a26cc Oct 07 13:45:39 crc kubenswrapper[4959]: I1007 13:45:39.297687 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:45:39 crc kubenswrapper[4959]: E1007 13:45:39.297974 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:45:40.297937304 +0000 UTC m=+22.381341639 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:45:39 crc kubenswrapper[4959]: I1007 13:45:39.298543 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:45:39 crc kubenswrapper[4959]: I1007 13:45:39.298593 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:45:39 crc kubenswrapper[4959]: I1007 13:45:39.298633 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:45:39 crc kubenswrapper[4959]: I1007 13:45:39.298667 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:45:39 crc kubenswrapper[4959]: E1007 13:45:39.298793 4959 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 13:45:39 crc kubenswrapper[4959]: E1007 13:45:39.298853 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 13:45:40.298842697 +0000 UTC m=+22.382247032 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 13:45:39 crc kubenswrapper[4959]: E1007 13:45:39.298910 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 13:45:39 crc kubenswrapper[4959]: E1007 13:45:39.298945 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 13:45:39 crc kubenswrapper[4959]: E1007 13:45:39.298993 4959 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:45:39 crc kubenswrapper[4959]: E1007 13:45:39.299036 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 13:45:39 crc kubenswrapper[4959]: E1007 13:45:39.299061 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 13:45:40.299041222 +0000 UTC m=+22.382445557 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:45:39 crc kubenswrapper[4959]: E1007 13:45:39.299078 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 13:45:39 crc kubenswrapper[4959]: E1007 13:45:39.299139 4959 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:45:39 crc kubenswrapper[4959]: E1007 13:45:39.299248 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 13:45:40.299217767 +0000 UTC m=+22.382622132 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:45:39 crc kubenswrapper[4959]: E1007 13:45:39.299387 4959 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 13:45:39 crc kubenswrapper[4959]: E1007 13:45:39.299444 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 13:45:40.299430122 +0000 UTC m=+22.382834457 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 13:45:39 crc kubenswrapper[4959]: I1007 13:45:39.652993 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:45:39 crc kubenswrapper[4959]: E1007 13:45:39.653311 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:45:39 crc kubenswrapper[4959]: I1007 13:45:39.787527 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"c3f72d6aa3541a3b359c575478b733a588d9765284fc2f2db84fd181311cbee5"} Oct 07 13:45:39 crc kubenswrapper[4959]: I1007 13:45:39.789723 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78"} Oct 07 13:45:39 crc kubenswrapper[4959]: I1007 13:45:39.789781 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"d75f5bbec9f3e389a92a5c78d116e35796e3e5f128b67ae4d609d44224621a0d"} Oct 07 13:45:39 crc kubenswrapper[4959]: I1007 13:45:39.792340 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649"} Oct 07 13:45:39 crc kubenswrapper[4959]: I1007 13:45:39.792376 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802"} Oct 07 13:45:39 crc kubenswrapper[4959]: I1007 13:45:39.792391 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"6647be4975a863bd92335c269adc5ea0936cd2ac8296b9add0b76702569a26cc"} Oct 07 13:45:39 crc kubenswrapper[4959]: I1007 13:45:39.809528 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:39Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:39 crc kubenswrapper[4959]: I1007 13:45:39.836357 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:39Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:39 crc kubenswrapper[4959]: I1007 13:45:39.854508 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:39Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:39 crc kubenswrapper[4959]: I1007 13:45:39.871019 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:39Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:39 crc kubenswrapper[4959]: I1007 13:45:39.888505 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:39Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:39 crc kubenswrapper[4959]: I1007 13:45:39.907333 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:39Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:39 crc kubenswrapper[4959]: I1007 13:45:39.925555 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:39Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:39 crc kubenswrapper[4959]: I1007 13:45:39.943035 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:39Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:39 crc kubenswrapper[4959]: I1007 13:45:39.970570 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:39Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:39 crc kubenswrapper[4959]: I1007 13:45:39.986779 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:39Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.000162 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:39Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.014273 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:40Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.028423 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:40Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.051873 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:40Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.308603 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:45:40 crc kubenswrapper[4959]: E1007 13:45:40.308821 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:45:42.308781215 +0000 UTC m=+24.392185590 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.308904 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.308954 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.308994 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.309036 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:45:40 crc kubenswrapper[4959]: E1007 13:45:40.309081 4959 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 13:45:40 crc kubenswrapper[4959]: E1007 13:45:40.309196 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 13:45:42.309175645 +0000 UTC m=+24.392579960 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 13:45:40 crc kubenswrapper[4959]: E1007 13:45:40.309197 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 13:45:40 crc kubenswrapper[4959]: E1007 13:45:40.309211 4959 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 13:45:40 crc kubenswrapper[4959]: E1007 13:45:40.309226 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 13:45:40 crc kubenswrapper[4959]: E1007 13:45:40.309249 4959 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:45:40 crc kubenswrapper[4959]: E1007 13:45:40.309275 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 13:45:40 crc kubenswrapper[4959]: E1007 13:45:40.309320 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 13:45:40 crc kubenswrapper[4959]: E1007 13:45:40.309288 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 13:45:42.309264787 +0000 UTC m=+24.392669142 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 13:45:40 crc kubenswrapper[4959]: E1007 13:45:40.309345 4959 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:45:40 crc kubenswrapper[4959]: E1007 13:45:40.309439 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 13:45:42.30937199 +0000 UTC m=+24.392776365 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:45:40 crc kubenswrapper[4959]: E1007 13:45:40.309525 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 13:45:42.309460282 +0000 UTC m=+24.392864757 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.653541 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.653594 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:45:40 crc kubenswrapper[4959]: E1007 13:45:40.653684 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:45:40 crc kubenswrapper[4959]: E1007 13:45:40.653793 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.660227 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.660906 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.661561 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.662138 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.662736 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.663245 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.663802 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.664364 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.664967 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.665474 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.665935 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.666644 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.667158 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.667671 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.668167 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.668676 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.671861 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.672260 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.672850 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.673882 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.674371 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.674924 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.675830 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.676684 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.677473 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.678073 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.679225 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.679672 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.680653 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.681090 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.681556 4959 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.682016 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.683753 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.684232 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.685132 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.686651 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.687323 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.688251 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.688824 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.689841 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.690429 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.691396 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.692004 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.692963 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.693431 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.694350 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.694947 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.696146 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.696593 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.697530 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.697984 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.698615 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.699606 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 07 13:45:40 crc kubenswrapper[4959]: I1007 13:45:40.700057 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 07 13:45:41 crc kubenswrapper[4959]: I1007 13:45:41.652964 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:45:41 crc kubenswrapper[4959]: E1007 13:45:41.653117 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.327841 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.327985 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:45:42 crc kubenswrapper[4959]: E1007 13:45:42.328263 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:45:46.327999194 +0000 UTC m=+28.411403549 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.328326 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.328377 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:45:42 crc kubenswrapper[4959]: E1007 13:45:42.328397 4959 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 13:45:42 crc kubenswrapper[4959]: E1007 13:45:42.328527 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 13:45:46.328455785 +0000 UTC m=+28.411860150 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 13:45:42 crc kubenswrapper[4959]: E1007 13:45:42.328621 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 13:45:42 crc kubenswrapper[4959]: E1007 13:45:42.328613 4959 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.328423 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:45:42 crc kubenswrapper[4959]: E1007 13:45:42.328683 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 13:45:42 crc kubenswrapper[4959]: E1007 13:45:42.328708 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 13:45:42 crc kubenswrapper[4959]: E1007 13:45:42.328731 4959 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:45:42 crc kubenswrapper[4959]: E1007 13:45:42.328648 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 13:45:42 crc kubenswrapper[4959]: E1007 13:45:42.328794 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 13:45:46.328749103 +0000 UTC m=+28.412153628 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 13:45:42 crc kubenswrapper[4959]: E1007 13:45:42.328834 4959 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:45:42 crc kubenswrapper[4959]: E1007 13:45:42.328843 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 13:45:46.328821394 +0000 UTC m=+28.412225979 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:45:42 crc kubenswrapper[4959]: E1007 13:45:42.328932 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 13:45:46.328905197 +0000 UTC m=+28.412309742 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.504561 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.521247 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.522827 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.523274 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:42Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.538506 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:42Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.561657 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:42Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.575522 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:42Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.596124 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:42Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.618933 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:42Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.643818 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:42Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.652951 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.652991 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:45:42 crc kubenswrapper[4959]: E1007 13:45:42.653157 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:45:42 crc kubenswrapper[4959]: E1007 13:45:42.653361 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.666911 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:42Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.686010 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:42Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.709419 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:42Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.729292 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:42Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.750274 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:42Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.787052 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:42Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.808694 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061"} Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.813370 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:42Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.834546 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:42Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.858548 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:42Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.875416 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:42Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.898804 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:42Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.917842 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:42Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.947262 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:42Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.967253 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:42Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:42 crc kubenswrapper[4959]: I1007 13:45:42.990493 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:42Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:43 crc kubenswrapper[4959]: I1007 13:45:43.012806 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:43Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:43 crc kubenswrapper[4959]: I1007 13:45:43.652491 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:45:43 crc kubenswrapper[4959]: E1007 13:45:43.652924 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.076224 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.078597 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.078648 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.078661 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.078727 4959 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.087015 4959 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.087320 4959 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.088583 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.088620 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.088630 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.088646 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.088659 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:44Z","lastTransitionTime":"2025-10-07T13:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:44 crc kubenswrapper[4959]: E1007 13:45:44.112570 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.115704 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.115736 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.115748 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.115763 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.115776 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:44Z","lastTransitionTime":"2025-10-07T13:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:44 crc kubenswrapper[4959]: E1007 13:45:44.131064 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.132634 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-j88pg"] Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.132936 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-j88pg" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.136060 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.137784 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.137933 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.141797 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.141834 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.141847 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.141864 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.141877 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:44Z","lastTransitionTime":"2025-10-07T13:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:44 crc kubenswrapper[4959]: E1007 13:45:44.160086 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.160692 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.164674 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.164716 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.164727 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.164744 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.164757 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:44Z","lastTransitionTime":"2025-10-07T13:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.176825 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: E1007 13:45:44.178817 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.182752 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.182789 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.182801 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.182817 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.182828 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:44Z","lastTransitionTime":"2025-10-07T13:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.193698 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: E1007 13:45:44.195719 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: E1007 13:45:44.195865 4959 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.197695 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.197731 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.197746 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.197764 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.197777 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:44Z","lastTransitionTime":"2025-10-07T13:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.206972 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.222432 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.253035 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.266032 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.278579 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/a0950962-bbba-48a7-8ac8-d6c83e4e61ac-hosts-file\") pod \"node-resolver-j88pg\" (UID: \"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\") " pod="openshift-dns/node-resolver-j88pg" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.278639 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgw42\" (UniqueName: \"kubernetes.io/projected/a0950962-bbba-48a7-8ac8-d6c83e4e61ac-kube-api-access-sgw42\") pod \"node-resolver-j88pg\" (UID: \"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\") " pod="openshift-dns/node-resolver-j88pg" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.298168 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.299542 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.299617 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.299631 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.299648 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.299661 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:44Z","lastTransitionTime":"2025-10-07T13:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.334479 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.379900 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/a0950962-bbba-48a7-8ac8-d6c83e4e61ac-hosts-file\") pod \"node-resolver-j88pg\" (UID: \"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\") " pod="openshift-dns/node-resolver-j88pg" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.379948 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgw42\" (UniqueName: \"kubernetes.io/projected/a0950962-bbba-48a7-8ac8-d6c83e4e61ac-kube-api-access-sgw42\") pod \"node-resolver-j88pg\" (UID: \"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\") " pod="openshift-dns/node-resolver-j88pg" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.380080 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/a0950962-bbba-48a7-8ac8-d6c83e4e61ac-hosts-file\") pod \"node-resolver-j88pg\" (UID: \"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\") " pod="openshift-dns/node-resolver-j88pg" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.397472 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgw42\" (UniqueName: \"kubernetes.io/projected/a0950962-bbba-48a7-8ac8-d6c83e4e61ac-kube-api-access-sgw42\") pod \"node-resolver-j88pg\" (UID: \"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\") " pod="openshift-dns/node-resolver-j88pg" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.402797 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.402831 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.402839 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.402852 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.402862 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:44Z","lastTransitionTime":"2025-10-07T13:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.448206 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-j88pg" Oct 07 13:45:44 crc kubenswrapper[4959]: W1007 13:45:44.462788 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda0950962_bbba_48a7_8ac8_d6c83e4e61ac.slice/crio-1324ce7293f3375c478feadc63fd78ed2dd820a7c1ad94b1c412ccbf98884526 WatchSource:0}: Error finding container 1324ce7293f3375c478feadc63fd78ed2dd820a7c1ad94b1c412ccbf98884526: Status 404 returned error can't find the container with id 1324ce7293f3375c478feadc63fd78ed2dd820a7c1ad94b1c412ccbf98884526 Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.505050 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.505087 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.505112 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.505127 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.505138 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:44Z","lastTransitionTime":"2025-10-07T13:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.547554 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-4mzl6"] Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.547941 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.550058 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.550216 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.550338 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.550640 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.553607 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.571486 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.596706 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.607325 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.607370 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.607383 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.607401 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.607415 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:44Z","lastTransitionTime":"2025-10-07T13:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.611577 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.622710 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.639165 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.651730 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.653497 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:45:44 crc kubenswrapper[4959]: E1007 13:45:44.653792 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.653928 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:45:44 crc kubenswrapper[4959]: E1007 13:45:44.654255 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.662082 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.673267 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.683020 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2003ec8f-74a0-47bc-8998-0326bfff6e7f-cni-binary-copy\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.683084 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-system-cni-dir\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.683193 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-multus-socket-dir-parent\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.683302 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-host-run-k8s-cni-cncf-io\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.683366 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-host-var-lib-cni-bin\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.683397 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-host-var-lib-cni-multus\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.683422 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-multus-conf-dir\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.683453 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-multus-cni-dir\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.683522 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-hostroot\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.683615 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/2003ec8f-74a0-47bc-8998-0326bfff6e7f-multus-daemon-config\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.683703 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-cnibin\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.683732 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-host-run-multus-certs\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.683768 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxqz7\" (UniqueName: \"kubernetes.io/projected/2003ec8f-74a0-47bc-8998-0326bfff6e7f-kube-api-access-jxqz7\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.683813 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-os-release\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.683856 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-etc-kubernetes\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.683900 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-host-run-netns\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.683937 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-host-var-lib-kubelet\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.687122 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.700023 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.710584 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.710624 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.710635 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.710653 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.710665 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:44Z","lastTransitionTime":"2025-10-07T13:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.784994 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-system-cni-dir\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785046 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2003ec8f-74a0-47bc-8998-0326bfff6e7f-cni-binary-copy\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785111 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-multus-socket-dir-parent\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785130 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-host-run-k8s-cni-cncf-io\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785145 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-host-var-lib-cni-bin\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785162 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-multus-conf-dir\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785212 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-multus-cni-dir\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785226 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-host-var-lib-cni-multus\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785239 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-hostroot\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785254 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/2003ec8f-74a0-47bc-8998-0326bfff6e7f-multus-daemon-config\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785276 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-cnibin\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785290 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-host-run-multus-certs\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785289 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-host-run-k8s-cni-cncf-io\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785336 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-system-cni-dir\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785381 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-multus-conf-dir\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785361 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-host-var-lib-cni-bin\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785616 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-multus-cni-dir\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785641 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-multus-socket-dir-parent\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785651 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-cnibin\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785308 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxqz7\" (UniqueName: \"kubernetes.io/projected/2003ec8f-74a0-47bc-8998-0326bfff6e7f-kube-api-access-jxqz7\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785676 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-host-run-multus-certs\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785692 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-etc-kubernetes\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785706 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-host-var-lib-cni-multus\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785715 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-os-release\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785730 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-etc-kubernetes\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785732 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-host-run-netns\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785753 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-host-run-netns\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785757 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-host-var-lib-kubelet\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785791 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-host-var-lib-kubelet\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785798 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-os-release\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.785819 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/2003ec8f-74a0-47bc-8998-0326bfff6e7f-hostroot\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.786374 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/2003ec8f-74a0-47bc-8998-0326bfff6e7f-multus-daemon-config\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.786642 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2003ec8f-74a0-47bc-8998-0326bfff6e7f-cni-binary-copy\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.802245 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxqz7\" (UniqueName: \"kubernetes.io/projected/2003ec8f-74a0-47bc-8998-0326bfff6e7f-kube-api-access-jxqz7\") pod \"multus-4mzl6\" (UID: \"2003ec8f-74a0-47bc-8998-0326bfff6e7f\") " pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.813389 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.813416 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.813426 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.813440 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.813452 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:44Z","lastTransitionTime":"2025-10-07T13:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.815838 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-j88pg" event={"ID":"a0950962-bbba-48a7-8ac8-d6c83e4e61ac","Type":"ContainerStarted","Data":"b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70"} Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.815890 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-j88pg" event={"ID":"a0950962-bbba-48a7-8ac8-d6c83e4e61ac","Type":"ContainerStarted","Data":"1324ce7293f3375c478feadc63fd78ed2dd820a7c1ad94b1c412ccbf98884526"} Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.842281 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.857611 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.860306 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-4mzl6" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.872216 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: W1007 13:45:44.872747 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2003ec8f_74a0_47bc_8998_0326bfff6e7f.slice/crio-becd7ffb2831e3048c16d73d45c6b55b9e448d78b3737e97651c7e70d53a39dd WatchSource:0}: Error finding container becd7ffb2831e3048c16d73d45c6b55b9e448d78b3737e97651c7e70d53a39dd: Status 404 returned error can't find the container with id becd7ffb2831e3048c16d73d45c6b55b9e448d78b3737e97651c7e70d53a39dd Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.887486 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.901384 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.915643 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.918973 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.919016 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.919033 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.919052 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.919066 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:44Z","lastTransitionTime":"2025-10-07T13:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.929680 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.934793 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5sfv5"] Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.935601 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-m5zzg"] Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.935853 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.935874 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-vxrtj"] Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.936070 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.936943 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.938990 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.939135 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.940239 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.941174 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.941308 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.941853 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.942086 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.942206 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.942609 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.943316 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.944461 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.949151 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.950316 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.950569 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.954056 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.968048 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.981588 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:44 crc kubenswrapper[4959]: I1007 13:45:44.995630 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:44Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.007682 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.017989 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.022588 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.022662 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.022678 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.022704 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.022723 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:45Z","lastTransitionTime":"2025-10-07T13:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.031158 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.044056 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.057117 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.070716 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.089220 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-run-netns\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.089263 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-run-ovn\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.089282 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-env-overrides\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.089318 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a-proxy-tls\") pod \"machine-config-daemon-m5zzg\" (UID: \"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\") " pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.089391 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-var-lib-openvswitch\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.089412 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-systemd-units\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.089436 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-cni-bin\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.089464 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/efa00eb4-4132-424a-99b9-4e57d30ce059-cnibin\") pod \"multus-additional-cni-plugins-vxrtj\" (UID: \"efa00eb4-4132-424a-99b9-4e57d30ce059\") " pod="openshift-multus/multus-additional-cni-plugins-vxrtj" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.089482 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-kubelet\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.089495 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-run-systemd\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.089515 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-log-socket\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.089624 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a-rootfs\") pod \"machine-config-daemon-m5zzg\" (UID: \"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\") " pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.089645 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.089708 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-ovnkube-config\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.089729 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-ovn-node-metrics-cert\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.089774 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/efa00eb4-4132-424a-99b9-4e57d30ce059-os-release\") pod \"multus-additional-cni-plugins-vxrtj\" (UID: \"efa00eb4-4132-424a-99b9-4e57d30ce059\") " pod="openshift-multus/multus-additional-cni-plugins-vxrtj" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.089868 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/efa00eb4-4132-424a-99b9-4e57d30ce059-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vxrtj\" (UID: \"efa00eb4-4132-424a-99b9-4e57d30ce059\") " pod="openshift-multus/multus-additional-cni-plugins-vxrtj" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.089968 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-node-log\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.090028 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a-mcd-auth-proxy-config\") pod \"machine-config-daemon-m5zzg\" (UID: \"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\") " pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.090121 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-run-ovn-kubernetes\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.090147 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-cni-netd\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.090185 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-ovnkube-script-lib\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.090249 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/efa00eb4-4132-424a-99b9-4e57d30ce059-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vxrtj\" (UID: \"efa00eb4-4132-424a-99b9-4e57d30ce059\") " pod="openshift-multus/multus-additional-cni-plugins-vxrtj" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.090302 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-slash\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.090342 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/efa00eb4-4132-424a-99b9-4e57d30ce059-system-cni-dir\") pod \"multus-additional-cni-plugins-vxrtj\" (UID: \"efa00eb4-4132-424a-99b9-4e57d30ce059\") " pod="openshift-multus/multus-additional-cni-plugins-vxrtj" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.090386 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xfdm\" (UniqueName: \"kubernetes.io/projected/efa00eb4-4132-424a-99b9-4e57d30ce059-kube-api-access-2xfdm\") pod \"multus-additional-cni-plugins-vxrtj\" (UID: \"efa00eb4-4132-424a-99b9-4e57d30ce059\") " pod="openshift-multus/multus-additional-cni-plugins-vxrtj" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.090422 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkdmv\" (UniqueName: \"kubernetes.io/projected/0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a-kube-api-access-lkdmv\") pod \"machine-config-daemon-m5zzg\" (UID: \"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\") " pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.090487 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-etc-openvswitch\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.090557 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-run-openvswitch\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.090600 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-969lv\" (UniqueName: \"kubernetes.io/projected/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-kube-api-access-969lv\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.090643 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/efa00eb4-4132-424a-99b9-4e57d30ce059-cni-binary-copy\") pod \"multus-additional-cni-plugins-vxrtj\" (UID: \"efa00eb4-4132-424a-99b9-4e57d30ce059\") " pod="openshift-multus/multus-additional-cni-plugins-vxrtj" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.091085 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.101680 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.107919 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.124560 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.124597 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.124606 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.124621 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.124632 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:45Z","lastTransitionTime":"2025-10-07T13:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.125220 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.162604 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.178608 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.190584 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191011 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-run-ovn-kubernetes\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191039 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-ovnkube-script-lib\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191056 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/efa00eb4-4132-424a-99b9-4e57d30ce059-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vxrtj\" (UID: \"efa00eb4-4132-424a-99b9-4e57d30ce059\") " pod="openshift-multus/multus-additional-cni-plugins-vxrtj" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191076 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-cni-netd\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191093 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-slash\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191127 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/efa00eb4-4132-424a-99b9-4e57d30ce059-system-cni-dir\") pod \"multus-additional-cni-plugins-vxrtj\" (UID: \"efa00eb4-4132-424a-99b9-4e57d30ce059\") " pod="openshift-multus/multus-additional-cni-plugins-vxrtj" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191148 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xfdm\" (UniqueName: \"kubernetes.io/projected/efa00eb4-4132-424a-99b9-4e57d30ce059-kube-api-access-2xfdm\") pod \"multus-additional-cni-plugins-vxrtj\" (UID: \"efa00eb4-4132-424a-99b9-4e57d30ce059\") " pod="openshift-multus/multus-additional-cni-plugins-vxrtj" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191165 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkdmv\" (UniqueName: \"kubernetes.io/projected/0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a-kube-api-access-lkdmv\") pod \"machine-config-daemon-m5zzg\" (UID: \"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\") " pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191190 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-etc-openvswitch\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191200 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-run-ovn-kubernetes\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191244 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-run-openvswitch\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191207 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-run-openvswitch\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191268 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-cni-netd\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191284 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-969lv\" (UniqueName: \"kubernetes.io/projected/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-kube-api-access-969lv\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191339 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/efa00eb4-4132-424a-99b9-4e57d30ce059-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vxrtj\" (UID: \"efa00eb4-4132-424a-99b9-4e57d30ce059\") " pod="openshift-multus/multus-additional-cni-plugins-vxrtj" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191382 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/efa00eb4-4132-424a-99b9-4e57d30ce059-cni-binary-copy\") pod \"multus-additional-cni-plugins-vxrtj\" (UID: \"efa00eb4-4132-424a-99b9-4e57d30ce059\") " pod="openshift-multus/multus-additional-cni-plugins-vxrtj" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191405 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-slash\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191431 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-run-netns\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191442 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-etc-openvswitch\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191465 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-run-ovn\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191476 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-run-netns\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191480 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/efa00eb4-4132-424a-99b9-4e57d30ce059-system-cni-dir\") pod \"multus-additional-cni-plugins-vxrtj\" (UID: \"efa00eb4-4132-424a-99b9-4e57d30ce059\") " pod="openshift-multus/multus-additional-cni-plugins-vxrtj" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191496 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-env-overrides\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191508 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-run-ovn\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191530 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a-proxy-tls\") pod \"machine-config-daemon-m5zzg\" (UID: \"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\") " pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191564 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-var-lib-openvswitch\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191599 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-systemd-units\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191627 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-cni-bin\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191661 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/efa00eb4-4132-424a-99b9-4e57d30ce059-cnibin\") pod \"multus-additional-cni-plugins-vxrtj\" (UID: \"efa00eb4-4132-424a-99b9-4e57d30ce059\") " pod="openshift-multus/multus-additional-cni-plugins-vxrtj" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191670 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-systemd-units\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191682 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-var-lib-openvswitch\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191706 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-cni-bin\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191714 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-run-systemd\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191735 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/efa00eb4-4132-424a-99b9-4e57d30ce059-cnibin\") pod \"multus-additional-cni-plugins-vxrtj\" (UID: \"efa00eb4-4132-424a-99b9-4e57d30ce059\") " pod="openshift-multus/multus-additional-cni-plugins-vxrtj" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191748 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-kubelet\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191764 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-run-systemd\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191778 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-log-socket\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191788 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-kubelet\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191808 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-log-socket\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191810 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a-rootfs\") pod \"machine-config-daemon-m5zzg\" (UID: \"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\") " pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191847 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-ovnkube-config\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191846 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a-rootfs\") pod \"machine-config-daemon-m5zzg\" (UID: \"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\") " pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191780 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-ovnkube-script-lib\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191863 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-ovn-node-metrics-cert\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191978 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/efa00eb4-4132-424a-99b9-4e57d30ce059-os-release\") pod \"multus-additional-cni-plugins-vxrtj\" (UID: \"efa00eb4-4132-424a-99b9-4e57d30ce059\") " pod="openshift-multus/multus-additional-cni-plugins-vxrtj" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191995 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/efa00eb4-4132-424a-99b9-4e57d30ce059-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vxrtj\" (UID: \"efa00eb4-4132-424a-99b9-4e57d30ce059\") " pod="openshift-multus/multus-additional-cni-plugins-vxrtj" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.191999 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-env-overrides\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.192012 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.192035 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.192036 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-node-log\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.192051 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/efa00eb4-4132-424a-99b9-4e57d30ce059-os-release\") pod \"multus-additional-cni-plugins-vxrtj\" (UID: \"efa00eb4-4132-424a-99b9-4e57d30ce059\") " pod="openshift-multus/multus-additional-cni-plugins-vxrtj" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.192059 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a-mcd-auth-proxy-config\") pod \"machine-config-daemon-m5zzg\" (UID: \"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\") " pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.192072 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/efa00eb4-4132-424a-99b9-4e57d30ce059-cni-binary-copy\") pod \"multus-additional-cni-plugins-vxrtj\" (UID: \"efa00eb4-4132-424a-99b9-4e57d30ce059\") " pod="openshift-multus/multus-additional-cni-plugins-vxrtj" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.192053 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-node-log\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.192522 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/efa00eb4-4132-424a-99b9-4e57d30ce059-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vxrtj\" (UID: \"efa00eb4-4132-424a-99b9-4e57d30ce059\") " pod="openshift-multus/multus-additional-cni-plugins-vxrtj" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.192641 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a-mcd-auth-proxy-config\") pod \"machine-config-daemon-m5zzg\" (UID: \"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\") " pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.192809 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-ovnkube-config\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.194831 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a-proxy-tls\") pod \"machine-config-daemon-m5zzg\" (UID: \"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\") " pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.194842 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-ovn-node-metrics-cert\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.204094 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.208067 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkdmv\" (UniqueName: \"kubernetes.io/projected/0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a-kube-api-access-lkdmv\") pod \"machine-config-daemon-m5zzg\" (UID: \"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\") " pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.209072 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xfdm\" (UniqueName: \"kubernetes.io/projected/efa00eb4-4132-424a-99b9-4e57d30ce059-kube-api-access-2xfdm\") pod \"multus-additional-cni-plugins-vxrtj\" (UID: \"efa00eb4-4132-424a-99b9-4e57d30ce059\") " pod="openshift-multus/multus-additional-cni-plugins-vxrtj" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.213643 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-969lv\" (UniqueName: \"kubernetes.io/projected/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-kube-api-access-969lv\") pod \"ovnkube-node-5sfv5\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.214498 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.226553 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.226593 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.226602 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.226617 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.226628 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:45Z","lastTransitionTime":"2025-10-07T13:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.228174 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.238516 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.249083 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.250921 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.261063 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.266732 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.266754 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: W1007 13:45:45.271072 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb79aa528_b2a2_4dff_b1d4_0e8dc918fc3b.slice/crio-0fadf3af91429ed4e0681e8592293f1bbc052706c27bf35d235e0778d9bce43c WatchSource:0}: Error finding container 0fadf3af91429ed4e0681e8592293f1bbc052706c27bf35d235e0778d9bce43c: Status 404 returned error can't find the container with id 0fadf3af91429ed4e0681e8592293f1bbc052706c27bf35d235e0778d9bce43c Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.273187 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" Oct 07 13:45:45 crc kubenswrapper[4959]: W1007 13:45:45.280021 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ac172b2_9dfc_446a_ba8f_6f1cb4df4c4a.slice/crio-e1701fc8407dd51f99eca100695e82937e29abafbb626f37b1581d4b7b038e9c WatchSource:0}: Error finding container e1701fc8407dd51f99eca100695e82937e29abafbb626f37b1581d4b7b038e9c: Status 404 returned error can't find the container with id e1701fc8407dd51f99eca100695e82937e29abafbb626f37b1581d4b7b038e9c Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.286628 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: W1007 13:45:45.292604 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podefa00eb4_4132_424a_99b9_4e57d30ce059.slice/crio-779b4437c98b545e4cc2ac1a4de98a680e5c5698b06c4f1990e3db8758b11f85 WatchSource:0}: Error finding container 779b4437c98b545e4cc2ac1a4de98a680e5c5698b06c4f1990e3db8758b11f85: Status 404 returned error can't find the container with id 779b4437c98b545e4cc2ac1a4de98a680e5c5698b06c4f1990e3db8758b11f85 Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.300092 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.317232 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.329431 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.329460 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.329470 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.329483 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.329492 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:45Z","lastTransitionTime":"2025-10-07T13:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.334227 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.352899 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.377575 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.395468 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.424479 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.434617 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.434670 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.434688 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.434711 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.434727 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:45Z","lastTransitionTime":"2025-10-07T13:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.438202 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.451392 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.465917 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.478202 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.491346 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.513513 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.533304 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.537320 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.537372 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.537413 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.537434 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.537482 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:45Z","lastTransitionTime":"2025-10-07T13:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.552808 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.567057 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.578719 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.612110 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.626025 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.639679 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.639722 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.639745 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.639771 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.639784 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:45Z","lastTransitionTime":"2025-10-07T13:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.640866 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.652630 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.652663 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: E1007 13:45:45.652738 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.665712 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.679588 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.741969 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.742012 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.742022 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.742035 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.742045 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:45Z","lastTransitionTime":"2025-10-07T13:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.820957 4959 generic.go:334] "Generic (PLEG): container finished" podID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerID="0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58" exitCode=0 Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.821035 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" event={"ID":"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b","Type":"ContainerDied","Data":"0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58"} Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.821299 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" event={"ID":"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b","Type":"ContainerStarted","Data":"0fadf3af91429ed4e0681e8592293f1bbc052706c27bf35d235e0778d9bce43c"} Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.823050 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" event={"ID":"efa00eb4-4132-424a-99b9-4e57d30ce059","Type":"ContainerStarted","Data":"5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a"} Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.823208 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" event={"ID":"efa00eb4-4132-424a-99b9-4e57d30ce059","Type":"ContainerStarted","Data":"779b4437c98b545e4cc2ac1a4de98a680e5c5698b06c4f1990e3db8758b11f85"} Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.825143 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerStarted","Data":"f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1"} Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.825221 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerStarted","Data":"bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f"} Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.825239 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerStarted","Data":"e1701fc8407dd51f99eca100695e82937e29abafbb626f37b1581d4b7b038e9c"} Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.826501 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4mzl6" event={"ID":"2003ec8f-74a0-47bc-8998-0326bfff6e7f","Type":"ContainerStarted","Data":"9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb"} Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.826590 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4mzl6" event={"ID":"2003ec8f-74a0-47bc-8998-0326bfff6e7f","Type":"ContainerStarted","Data":"becd7ffb2831e3048c16d73d45c6b55b9e448d78b3737e97651c7e70d53a39dd"} Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.844916 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.844965 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.844975 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.844994 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.845008 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:45Z","lastTransitionTime":"2025-10-07T13:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.847681 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.863939 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.878946 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.892400 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.913214 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.933000 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.949787 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.949841 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.949854 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.949874 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.949889 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:45Z","lastTransitionTime":"2025-10-07T13:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.952452 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.965324 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:45 crc kubenswrapper[4959]: I1007 13:45:45.982621 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.000036 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.014066 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:46Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.030801 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:46Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.046076 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:46Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.053274 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.053335 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.053354 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.053426 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.053452 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:46Z","lastTransitionTime":"2025-10-07T13:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.071564 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:46Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.107294 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:46Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.144710 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:46Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.156916 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.157000 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.157022 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.157049 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.157068 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:46Z","lastTransitionTime":"2025-10-07T13:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.188381 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:46Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.229470 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:46Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.260546 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.260591 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.260604 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.260626 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.260640 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:46Z","lastTransitionTime":"2025-10-07T13:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.267441 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:46Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.306086 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:46Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.344628 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:46Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.363298 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.363338 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.363347 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.363361 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.363371 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:46Z","lastTransitionTime":"2025-10-07T13:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.387198 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:46Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.403782 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:45:46 crc kubenswrapper[4959]: E1007 13:45:46.403962 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:45:54.40393086 +0000 UTC m=+36.487335195 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.404076 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.404177 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:45:46 crc kubenswrapper[4959]: E1007 13:45:46.404239 4959 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 13:45:46 crc kubenswrapper[4959]: E1007 13:45:46.404310 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 13:45:54.40429397 +0000 UTC m=+36.487698295 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.404237 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:45:46 crc kubenswrapper[4959]: E1007 13:45:46.404357 4959 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.404369 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:45:46 crc kubenswrapper[4959]: E1007 13:45:46.404382 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 13:45:46 crc kubenswrapper[4959]: E1007 13:45:46.404403 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 13:45:46 crc kubenswrapper[4959]: E1007 13:45:46.404417 4959 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:45:46 crc kubenswrapper[4959]: E1007 13:45:46.404465 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 13:45:54.404437453 +0000 UTC m=+36.487841978 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 13:45:46 crc kubenswrapper[4959]: E1007 13:45:46.404499 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 13:45:54.404485005 +0000 UTC m=+36.487889590 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:45:46 crc kubenswrapper[4959]: E1007 13:45:46.404508 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 13:45:46 crc kubenswrapper[4959]: E1007 13:45:46.404524 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 13:45:46 crc kubenswrapper[4959]: E1007 13:45:46.404535 4959 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:45:46 crc kubenswrapper[4959]: E1007 13:45:46.404560 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 13:45:54.404553166 +0000 UTC m=+36.487957491 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.436818 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:46Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.466520 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:46Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.466848 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.466940 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.466966 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.466992 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.467012 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:46Z","lastTransitionTime":"2025-10-07T13:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.504615 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:46Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.541946 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:46Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.570829 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.570897 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.570918 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.570949 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.570968 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:46Z","lastTransitionTime":"2025-10-07T13:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.582947 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:46Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.631196 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:46Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.652540 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:45:46 crc kubenswrapper[4959]: E1007 13:45:46.652726 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.653251 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:45:46 crc kubenswrapper[4959]: E1007 13:45:46.653353 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.673239 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.673276 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.673285 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.673300 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.673309 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:46Z","lastTransitionTime":"2025-10-07T13:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.776438 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.776508 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.776530 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.776560 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.776585 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:46Z","lastTransitionTime":"2025-10-07T13:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.832990 4959 generic.go:334] "Generic (PLEG): container finished" podID="efa00eb4-4132-424a-99b9-4e57d30ce059" containerID="5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a" exitCode=0 Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.833034 4959 generic.go:334] "Generic (PLEG): container finished" podID="efa00eb4-4132-424a-99b9-4e57d30ce059" containerID="a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839" exitCode=0 Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.833093 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" event={"ID":"efa00eb4-4132-424a-99b9-4e57d30ce059","Type":"ContainerDied","Data":"5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a"} Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.833145 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" event={"ID":"efa00eb4-4132-424a-99b9-4e57d30ce059","Type":"ContainerDied","Data":"a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839"} Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.839960 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" event={"ID":"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b","Type":"ContainerStarted","Data":"bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c"} Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.840024 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" event={"ID":"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b","Type":"ContainerStarted","Data":"7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5"} Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.840047 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" event={"ID":"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b","Type":"ContainerStarted","Data":"d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3"} Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.840069 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" event={"ID":"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b","Type":"ContainerStarted","Data":"f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a"} Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.840087 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" event={"ID":"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b","Type":"ContainerStarted","Data":"1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f"} Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.855887 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:46Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.876268 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:46Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.882892 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.882952 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.882970 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.882996 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.883015 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:46Z","lastTransitionTime":"2025-10-07T13:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.888835 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:46Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.910497 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:46Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.924807 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:46Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.942608 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:46Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.956741 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:46Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.969486 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:46Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.986036 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.986193 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.986224 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.986249 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.986264 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:46Z","lastTransitionTime":"2025-10-07T13:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:46 crc kubenswrapper[4959]: I1007 13:45:46.986965 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:46Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.024343 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:47Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.066577 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:47Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.088556 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.088592 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.088603 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.088621 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.088633 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:47Z","lastTransitionTime":"2025-10-07T13:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.104165 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:47Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.145670 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:47Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.204443 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:47Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.206435 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.206462 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.206492 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.206508 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.206517 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:47Z","lastTransitionTime":"2025-10-07T13:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.309070 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.309112 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.309120 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.309133 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.309141 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:47Z","lastTransitionTime":"2025-10-07T13:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.411163 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.411210 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.411221 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.411236 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.411251 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:47Z","lastTransitionTime":"2025-10-07T13:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.514658 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.514720 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.514739 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.514766 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.514785 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:47Z","lastTransitionTime":"2025-10-07T13:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.591489 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-km89w"] Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.592354 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-km89w" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.594843 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.595322 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.595378 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.595498 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.618190 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.618250 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.618263 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.618294 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.618308 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:47Z","lastTransitionTime":"2025-10-07T13:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.626322 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:47Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.644396 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:47Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.652828 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:45:47 crc kubenswrapper[4959]: E1007 13:45:47.653021 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.663311 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:47Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.678400 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:47Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.691573 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:47Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.718410 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:47Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.719493 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s45mf\" (UniqueName: \"kubernetes.io/projected/1c27a0cb-3867-4c31-b438-c34570f7e1cd-kube-api-access-s45mf\") pod \"node-ca-km89w\" (UID: \"1c27a0cb-3867-4c31-b438-c34570f7e1cd\") " pod="openshift-image-registry/node-ca-km89w" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.719534 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/1c27a0cb-3867-4c31-b438-c34570f7e1cd-serviceca\") pod \"node-ca-km89w\" (UID: \"1c27a0cb-3867-4c31-b438-c34570f7e1cd\") " pod="openshift-image-registry/node-ca-km89w" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.719561 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1c27a0cb-3867-4c31-b438-c34570f7e1cd-host\") pod \"node-ca-km89w\" (UID: \"1c27a0cb-3867-4c31-b438-c34570f7e1cd\") " pod="openshift-image-registry/node-ca-km89w" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.720604 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.720628 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.720636 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.720650 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.720660 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:47Z","lastTransitionTime":"2025-10-07T13:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.733271 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:47Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.748605 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:47Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.762880 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:47Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.781745 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:47Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.798534 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:47Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.814019 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:47Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.820455 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1c27a0cb-3867-4c31-b438-c34570f7e1cd-host\") pod \"node-ca-km89w\" (UID: \"1c27a0cb-3867-4c31-b438-c34570f7e1cd\") " pod="openshift-image-registry/node-ca-km89w" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.820577 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s45mf\" (UniqueName: \"kubernetes.io/projected/1c27a0cb-3867-4c31-b438-c34570f7e1cd-kube-api-access-s45mf\") pod \"node-ca-km89w\" (UID: \"1c27a0cb-3867-4c31-b438-c34570f7e1cd\") " pod="openshift-image-registry/node-ca-km89w" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.820624 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1c27a0cb-3867-4c31-b438-c34570f7e1cd-host\") pod \"node-ca-km89w\" (UID: \"1c27a0cb-3867-4c31-b438-c34570f7e1cd\") " pod="openshift-image-registry/node-ca-km89w" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.820652 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/1c27a0cb-3867-4c31-b438-c34570f7e1cd-serviceca\") pod \"node-ca-km89w\" (UID: \"1c27a0cb-3867-4c31-b438-c34570f7e1cd\") " pod="openshift-image-registry/node-ca-km89w" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.822688 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.822731 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.822740 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.822755 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.822765 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:47Z","lastTransitionTime":"2025-10-07T13:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.822949 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/1c27a0cb-3867-4c31-b438-c34570f7e1cd-serviceca\") pod \"node-ca-km89w\" (UID: \"1c27a0cb-3867-4c31-b438-c34570f7e1cd\") " pod="openshift-image-registry/node-ca-km89w" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.835005 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:47Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.845136 4959 generic.go:334] "Generic (PLEG): container finished" podID="efa00eb4-4132-424a-99b9-4e57d30ce059" containerID="e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d" exitCode=0 Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.845219 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" event={"ID":"efa00eb4-4132-424a-99b9-4e57d30ce059","Type":"ContainerDied","Data":"e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d"} Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.849586 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" event={"ID":"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b","Type":"ContainerStarted","Data":"3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b"} Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.853648 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:47Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.857994 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s45mf\" (UniqueName: \"kubernetes.io/projected/1c27a0cb-3867-4c31-b438-c34570f7e1cd-kube-api-access-s45mf\") pod \"node-ca-km89w\" (UID: \"1c27a0cb-3867-4c31-b438-c34570f7e1cd\") " pod="openshift-image-registry/node-ca-km89w" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.890156 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:47Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.915013 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-km89w" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.922340 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:47Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.925154 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.925198 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.925210 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.925233 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.925248 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:47Z","lastTransitionTime":"2025-10-07T13:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:47 crc kubenswrapper[4959]: W1007 13:45:47.929038 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c27a0cb_3867_4c31_b438_c34570f7e1cd.slice/crio-00f9ef3098d515505bdf4047d7165ab393cc517ccbcaa406e2005464f26918d5 WatchSource:0}: Error finding container 00f9ef3098d515505bdf4047d7165ab393cc517ccbcaa406e2005464f26918d5: Status 404 returned error can't find the container with id 00f9ef3098d515505bdf4047d7165ab393cc517ccbcaa406e2005464f26918d5 Oct 07 13:45:47 crc kubenswrapper[4959]: I1007 13:45:47.964449 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:47Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.003570 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.032779 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.033263 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.033351 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.033379 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.033395 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:48Z","lastTransitionTime":"2025-10-07T13:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.043305 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.084094 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.131425 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.137206 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.137258 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.137275 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.137326 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.137344 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:48Z","lastTransitionTime":"2025-10-07T13:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.164556 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.203733 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.241083 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.241160 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.241171 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.241194 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.241204 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:48Z","lastTransitionTime":"2025-10-07T13:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.243007 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.289558 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.332027 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.343733 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.343759 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.343768 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.343782 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.343793 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:48Z","lastTransitionTime":"2025-10-07T13:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.374768 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.410885 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.449559 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.449605 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.449614 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.449628 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.449669 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:48Z","lastTransitionTime":"2025-10-07T13:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.452723 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.488756 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.553311 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.553388 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.553411 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.553444 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.553471 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:48Z","lastTransitionTime":"2025-10-07T13:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.653382 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.653409 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:45:48 crc kubenswrapper[4959]: E1007 13:45:48.653602 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:45:48 crc kubenswrapper[4959]: E1007 13:45:48.653750 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.655858 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.655916 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.655937 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.655961 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.655981 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:48Z","lastTransitionTime":"2025-10-07T13:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.683803 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.702762 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.719307 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.734306 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.753523 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.761925 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.761963 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.761976 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.761991 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.762002 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:48Z","lastTransitionTime":"2025-10-07T13:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.770490 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.788397 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.809660 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.846719 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.856284 4959 generic.go:334] "Generic (PLEG): container finished" podID="efa00eb4-4132-424a-99b9-4e57d30ce059" containerID="d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0" exitCode=0 Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.856387 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" event={"ID":"efa00eb4-4132-424a-99b9-4e57d30ce059","Type":"ContainerDied","Data":"d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0"} Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.858981 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-km89w" event={"ID":"1c27a0cb-3867-4c31-b438-c34570f7e1cd","Type":"ContainerStarted","Data":"e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8"} Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.859017 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-km89w" event={"ID":"1c27a0cb-3867-4c31-b438-c34570f7e1cd","Type":"ContainerStarted","Data":"00f9ef3098d515505bdf4047d7165ab393cc517ccbcaa406e2005464f26918d5"} Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.864912 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.865008 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.865025 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.865041 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.865054 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:48Z","lastTransitionTime":"2025-10-07T13:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.893491 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.922403 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.964037 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.967714 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.967774 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.967791 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.967814 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:48 crc kubenswrapper[4959]: I1007 13:45:48.967831 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:48Z","lastTransitionTime":"2025-10-07T13:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.006822 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:49Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.045246 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:49Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.069510 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.069539 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.069547 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.069561 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.069571 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:49Z","lastTransitionTime":"2025-10-07T13:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.087876 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:49Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.122828 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:49Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.174235 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.174297 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.174316 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.174343 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.174361 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:49Z","lastTransitionTime":"2025-10-07T13:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.186151 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:49Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.206995 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:49Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.243451 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:49Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.278479 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.278824 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.278957 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.279045 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.279189 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:49Z","lastTransitionTime":"2025-10-07T13:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.284337 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:49Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.328320 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:49Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.367066 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:49Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.381960 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.382031 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.382054 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.382084 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.382137 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:49Z","lastTransitionTime":"2025-10-07T13:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.412988 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:49Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.444880 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:49Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.485949 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:49Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.486891 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.486962 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.486980 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.487007 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.487026 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:49Z","lastTransitionTime":"2025-10-07T13:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.526042 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:49Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.566525 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:49Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.590640 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.591008 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.591028 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.591054 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.591073 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:49Z","lastTransitionTime":"2025-10-07T13:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.614870 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:49Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.645885 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:49Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.653323 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:45:49 crc kubenswrapper[4959]: E1007 13:45:49.653517 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.687830 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:49Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.693506 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.693567 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.693585 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.693609 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.693627 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:49Z","lastTransitionTime":"2025-10-07T13:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.797173 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.797249 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.797271 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.797300 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.797333 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:49Z","lastTransitionTime":"2025-10-07T13:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.868203 4959 generic.go:334] "Generic (PLEG): container finished" podID="efa00eb4-4132-424a-99b9-4e57d30ce059" containerID="7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee" exitCode=0 Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.868372 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" event={"ID":"efa00eb4-4132-424a-99b9-4e57d30ce059","Type":"ContainerDied","Data":"7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee"} Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.879457 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" event={"ID":"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b","Type":"ContainerStarted","Data":"ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981"} Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.888391 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:49Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.900865 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.900953 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.901011 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.901042 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.901061 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:49Z","lastTransitionTime":"2025-10-07T13:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.911236 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:49Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.933355 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:49Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.956792 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:49Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.971020 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:49Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:49 crc kubenswrapper[4959]: I1007 13:45:49.990939 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:49Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.004132 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.004173 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.004188 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.004208 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.004226 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:50Z","lastTransitionTime":"2025-10-07T13:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.007012 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:50Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.022683 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:50Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.053659 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:50Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.086046 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:50Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.107081 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.107408 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.107504 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.107627 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.107737 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:50Z","lastTransitionTime":"2025-10-07T13:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.126221 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:50Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.163485 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:50Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.204924 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:50Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.210728 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.210770 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.210782 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.210798 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.210809 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:50Z","lastTransitionTime":"2025-10-07T13:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.250912 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:50Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.283951 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:50Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.319304 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.319372 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.319395 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.319424 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.319445 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:50Z","lastTransitionTime":"2025-10-07T13:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.424035 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.424421 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.424581 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.424831 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.425129 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:50Z","lastTransitionTime":"2025-10-07T13:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.527844 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.527899 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.527935 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.527969 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.528154 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:50Z","lastTransitionTime":"2025-10-07T13:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.631073 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.631129 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.631141 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.631158 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.631170 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:50Z","lastTransitionTime":"2025-10-07T13:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.652728 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:45:50 crc kubenswrapper[4959]: E1007 13:45:50.652909 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.652970 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:45:50 crc kubenswrapper[4959]: E1007 13:45:50.653256 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.733458 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.733528 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.733546 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.733569 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.733585 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:50Z","lastTransitionTime":"2025-10-07T13:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.835277 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.835337 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.835356 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.835381 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.835399 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:50Z","lastTransitionTime":"2025-10-07T13:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.884322 4959 generic.go:334] "Generic (PLEG): container finished" podID="efa00eb4-4132-424a-99b9-4e57d30ce059" containerID="ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0" exitCode=0 Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.884364 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" event={"ID":"efa00eb4-4132-424a-99b9-4e57d30ce059","Type":"ContainerDied","Data":"ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0"} Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.912200 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:50Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.937848 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.937914 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.937930 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.937957 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.937980 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:50Z","lastTransitionTime":"2025-10-07T13:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.940625 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:50Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.954888 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:50Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.968064 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:50Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.977867 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:50Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:50 crc kubenswrapper[4959]: I1007 13:45:50.988654 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:50Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.009150 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:51Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.023371 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:51Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.034991 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:51Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.039852 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.039882 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.039891 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.039903 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.039912 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:51Z","lastTransitionTime":"2025-10-07T13:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.048892 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:51Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.061443 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:51Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.075318 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:51Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.085034 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:51Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.100143 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:51Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.111592 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:51Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.141925 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.141972 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.141983 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.142000 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.142012 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:51Z","lastTransitionTime":"2025-10-07T13:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.248771 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.248830 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.248849 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.248872 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.248891 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:51Z","lastTransitionTime":"2025-10-07T13:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.353397 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.353454 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.353471 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.353493 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.353510 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:51Z","lastTransitionTime":"2025-10-07T13:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.458641 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.459156 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.459183 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.459218 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.459241 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:51Z","lastTransitionTime":"2025-10-07T13:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.561660 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.561702 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.561715 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.561734 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.561749 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:51Z","lastTransitionTime":"2025-10-07T13:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.653131 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:45:51 crc kubenswrapper[4959]: E1007 13:45:51.653261 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.664224 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.664247 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.664256 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.664268 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.664277 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:51Z","lastTransitionTime":"2025-10-07T13:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.766859 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.766913 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.766931 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.766954 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.766971 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:51Z","lastTransitionTime":"2025-10-07T13:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.869335 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.869397 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.869418 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.869473 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.869495 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:51Z","lastTransitionTime":"2025-10-07T13:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.893042 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" event={"ID":"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b","Type":"ContainerStarted","Data":"ee769877d15b988363ec0512da1e1fb400821b9f241038a67fac9b334b6340c7"} Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.893417 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.899211 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" event={"ID":"efa00eb4-4132-424a-99b9-4e57d30ce059","Type":"ContainerStarted","Data":"c1cbef019f0640b5d51b398d1fc7166c1e253fd33f78f0e50f9754bb1187a6d0"} Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.916846 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:51Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.927299 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.928337 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:51Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.946298 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:51Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.960510 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:51Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.971769 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.971947 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.972003 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.972076 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.972173 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:51Z","lastTransitionTime":"2025-10-07T13:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.972348 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:51Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:51 crc kubenswrapper[4959]: I1007 13:45:51.990429 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:51Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.008133 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:52Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.020852 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:52Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.034002 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:52Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.059791 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee769877d15b988363ec0512da1e1fb400821b9f241038a67fac9b334b6340c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:52Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.073629 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.073671 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.073688 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.073710 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.073745 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:52Z","lastTransitionTime":"2025-10-07T13:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.085084 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:52Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.102587 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:52Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.119179 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:52Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.130201 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:52Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.145268 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:52Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.155896 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:52Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.169067 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:52Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.176994 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.177166 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.177225 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.177286 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.177348 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:52Z","lastTransitionTime":"2025-10-07T13:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.185932 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:52Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.198214 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:52Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.232945 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:52Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.258897 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1cbef019f0640b5d51b398d1fc7166c1e253fd33f78f0e50f9754bb1187a6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:52Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.272594 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:52Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.280063 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.280234 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.280323 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.280384 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.280456 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:52Z","lastTransitionTime":"2025-10-07T13:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.286267 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:52Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.296955 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:52Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.312431 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee769877d15b988363ec0512da1e1fb400821b9f241038a67fac9b334b6340c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:52Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.328020 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:52Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.341043 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:52Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.352346 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:52Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.360577 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:52Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.371836 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:52Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.382546 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.382671 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.382755 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.382851 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.382931 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:52Z","lastTransitionTime":"2025-10-07T13:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.484976 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.485019 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.485078 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.485122 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.485140 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:52Z","lastTransitionTime":"2025-10-07T13:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.587669 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.587713 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.587725 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.587742 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.587757 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:52Z","lastTransitionTime":"2025-10-07T13:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.653019 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:45:52 crc kubenswrapper[4959]: E1007 13:45:52.653241 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.653316 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:45:52 crc kubenswrapper[4959]: E1007 13:45:52.653550 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.690918 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.691195 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.691305 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.691457 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.691546 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:52Z","lastTransitionTime":"2025-10-07T13:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.794796 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.794843 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.794856 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.794879 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.794892 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:52Z","lastTransitionTime":"2025-10-07T13:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.898465 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.898897 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.899041 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.899222 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.899408 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:52Z","lastTransitionTime":"2025-10-07T13:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.903145 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.903341 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:52 crc kubenswrapper[4959]: I1007 13:45:52.985546 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.001808 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.001850 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.001862 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.001880 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.001893 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:53Z","lastTransitionTime":"2025-10-07T13:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.005908 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:53Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.020358 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:53Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.032865 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:53Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.055523 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee769877d15b988363ec0512da1e1fb400821b9f241038a67fac9b334b6340c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:53Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.077998 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:53Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.097472 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:53Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.104601 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.104642 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.104668 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.104699 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.104724 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:53Z","lastTransitionTime":"2025-10-07T13:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.117248 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:53Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.133269 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:53Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.145458 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:53Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.161752 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1cbef019f0640b5d51b398d1fc7166c1e253fd33f78f0e50f9754bb1187a6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:53Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.171936 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:53Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.186599 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:53Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.200581 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:53Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.206807 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.206838 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.206850 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.206866 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.206880 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:53Z","lastTransitionTime":"2025-10-07T13:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.215581 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:53Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.230297 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:53Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.309830 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.309863 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.309875 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.309892 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.309906 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:53Z","lastTransitionTime":"2025-10-07T13:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.413256 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.413349 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.413365 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.413385 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.413402 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:53Z","lastTransitionTime":"2025-10-07T13:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.517222 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.517287 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.517307 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.517369 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.517389 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:53Z","lastTransitionTime":"2025-10-07T13:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.621034 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.621152 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.621161 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.621174 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.621182 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:53Z","lastTransitionTime":"2025-10-07T13:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.653468 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:45:53 crc kubenswrapper[4959]: E1007 13:45:53.653602 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.722755 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.722782 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.722804 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.722816 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.722825 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:53Z","lastTransitionTime":"2025-10-07T13:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.824802 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.824832 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.824840 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.824854 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.824864 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:53Z","lastTransitionTime":"2025-10-07T13:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.927250 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.927289 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.927300 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.927316 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:53 crc kubenswrapper[4959]: I1007 13:45:53.927332 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:53Z","lastTransitionTime":"2025-10-07T13:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.029133 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.029161 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.029170 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.029182 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.029192 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:54Z","lastTransitionTime":"2025-10-07T13:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.131263 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.131318 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.131335 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.131357 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.131374 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:54Z","lastTransitionTime":"2025-10-07T13:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.234464 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.234541 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.234564 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.234596 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.234618 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:54Z","lastTransitionTime":"2025-10-07T13:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.337958 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.338025 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.338048 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.338079 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.338134 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:54Z","lastTransitionTime":"2025-10-07T13:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.440747 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.440806 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.440823 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.440847 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.440867 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:54Z","lastTransitionTime":"2025-10-07T13:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.494658 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.494746 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.494772 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.494804 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.494827 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:54Z","lastTransitionTime":"2025-10-07T13:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.498245 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:45:54 crc kubenswrapper[4959]: E1007 13:45:54.498378 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:46:10.498351098 +0000 UTC m=+52.581755453 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.498439 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.498485 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.498531 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.498576 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:45:54 crc kubenswrapper[4959]: E1007 13:45:54.498625 4959 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 13:45:54 crc kubenswrapper[4959]: E1007 13:45:54.498711 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 13:46:10.498688747 +0000 UTC m=+52.582093102 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 13:45:54 crc kubenswrapper[4959]: E1007 13:45:54.498716 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 13:45:54 crc kubenswrapper[4959]: E1007 13:45:54.498739 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 13:45:54 crc kubenswrapper[4959]: E1007 13:45:54.498758 4959 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:45:54 crc kubenswrapper[4959]: E1007 13:45:54.498756 4959 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 13:45:54 crc kubenswrapper[4959]: E1007 13:45:54.498816 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 13:46:10.49880138 +0000 UTC m=+52.582205735 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:45:54 crc kubenswrapper[4959]: E1007 13:45:54.498866 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 13:46:10.498839121 +0000 UTC m=+52.582243486 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 13:45:54 crc kubenswrapper[4959]: E1007 13:45:54.498758 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 13:45:54 crc kubenswrapper[4959]: E1007 13:45:54.498906 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 13:45:54 crc kubenswrapper[4959]: E1007 13:45:54.498926 4959 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:45:54 crc kubenswrapper[4959]: E1007 13:45:54.498987 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 13:46:10.498974314 +0000 UTC m=+52.582378669 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:45:54 crc kubenswrapper[4959]: E1007 13:45:54.516517 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:54Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.521617 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.521696 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.521722 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.521755 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.521782 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:54Z","lastTransitionTime":"2025-10-07T13:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:54 crc kubenswrapper[4959]: E1007 13:45:54.537062 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:54Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.543146 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.543200 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.543214 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.543234 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.543247 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:54Z","lastTransitionTime":"2025-10-07T13:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:54 crc kubenswrapper[4959]: E1007 13:45:54.557439 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:54Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.564557 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.564650 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.564669 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.564719 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.564737 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:54Z","lastTransitionTime":"2025-10-07T13:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:54 crc kubenswrapper[4959]: E1007 13:45:54.587612 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:54Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.591689 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.591733 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.591749 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.591771 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.591788 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:54Z","lastTransitionTime":"2025-10-07T13:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:54 crc kubenswrapper[4959]: E1007 13:45:54.608556 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:54Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:54 crc kubenswrapper[4959]: E1007 13:45:54.608779 4959 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.610437 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.610482 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.610504 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.610527 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.610547 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:54Z","lastTransitionTime":"2025-10-07T13:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.652691 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:45:54 crc kubenswrapper[4959]: E1007 13:45:54.652854 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.653402 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:45:54 crc kubenswrapper[4959]: E1007 13:45:54.653509 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.713545 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.713868 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.713884 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.713903 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.713917 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:54Z","lastTransitionTime":"2025-10-07T13:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.816052 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.816130 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.816143 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.816163 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.816176 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:54Z","lastTransitionTime":"2025-10-07T13:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.910201 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5sfv5_b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b/ovnkube-controller/0.log" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.914307 4959 generic.go:334] "Generic (PLEG): container finished" podID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerID="ee769877d15b988363ec0512da1e1fb400821b9f241038a67fac9b334b6340c7" exitCode=1 Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.914351 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" event={"ID":"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b","Type":"ContainerDied","Data":"ee769877d15b988363ec0512da1e1fb400821b9f241038a67fac9b334b6340c7"} Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.915126 4959 scope.go:117] "RemoveContainer" containerID="ee769877d15b988363ec0512da1e1fb400821b9f241038a67fac9b334b6340c7" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.918272 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.918306 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.918319 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.918334 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.918347 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:54Z","lastTransitionTime":"2025-10-07T13:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.943089 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:54Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.958980 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:54Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.973263 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:54Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.983153 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:54Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:54 crc kubenswrapper[4959]: I1007 13:45:54.993373 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:54Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.002816 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:55Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.020342 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.020477 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.020557 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.020677 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.020756 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:55Z","lastTransitionTime":"2025-10-07T13:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.020719 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:55Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.030524 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:55Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.040929 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:55Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.053690 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:55Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.069734 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1cbef019f0640b5d51b398d1fc7166c1e253fd33f78f0e50f9754bb1187a6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:55Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.085981 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:55Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.100084 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:55Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.112324 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:55Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.122658 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.122687 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.122694 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.122706 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.122715 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:55Z","lastTransitionTime":"2025-10-07T13:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.130942 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee769877d15b988363ec0512da1e1fb400821b9f241038a67fac9b334b6340c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee769877d15b988363ec0512da1e1fb400821b9f241038a67fac9b334b6340c7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"message\\\":\\\" 6251 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1007 13:45:54.475870 6251 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1007 13:45:54.475918 6251 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1007 13:45:54.475944 6251 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1007 13:45:54.475980 6251 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1007 13:45:54.475989 6251 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1007 13:45:54.476011 6251 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 13:45:54.476021 6251 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1007 13:45:54.476079 6251 factory.go:656] Stopping watch factory\\\\nI1007 13:45:54.476134 6251 ovnkube.go:599] Stopped ovnkube\\\\nI1007 13:45:54.476179 6251 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1007 13:45:54.476195 6251 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1007 13:45:54.476209 6251 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1007 13:45:54.476223 6251 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1007 13:45:54.476238 6251 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1007 13:45:54.476252 6251 handler.go:208] Removed *v1.Node event handler 2\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:55Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.224882 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.224918 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.224926 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.224942 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.224950 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:55Z","lastTransitionTime":"2025-10-07T13:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.326984 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.327025 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.327034 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.327048 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.327057 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:55Z","lastTransitionTime":"2025-10-07T13:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.429884 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.429926 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.429939 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.429965 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.429977 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:55Z","lastTransitionTime":"2025-10-07T13:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.532603 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.532642 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.532651 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.532664 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.532673 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:55Z","lastTransitionTime":"2025-10-07T13:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.635028 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.635064 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.635076 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.635089 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.635114 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:55Z","lastTransitionTime":"2025-10-07T13:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.653349 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:45:55 crc kubenswrapper[4959]: E1007 13:45:55.653490 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.737554 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.737595 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.737607 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.737623 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.737644 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:55Z","lastTransitionTime":"2025-10-07T13:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.840706 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.840761 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.840778 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.840800 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.840817 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:55Z","lastTransitionTime":"2025-10-07T13:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.921581 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5sfv5_b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b/ovnkube-controller/1.log" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.925732 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5sfv5_b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b/ovnkube-controller/0.log" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.930720 4959 generic.go:334] "Generic (PLEG): container finished" podID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerID="63f045b52f496803aa5e9e76437ab3594521f8d8772b5b88332ebc178da54515" exitCode=1 Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.930781 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" event={"ID":"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b","Type":"ContainerDied","Data":"63f045b52f496803aa5e9e76437ab3594521f8d8772b5b88332ebc178da54515"} Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.930885 4959 scope.go:117] "RemoveContainer" containerID="ee769877d15b988363ec0512da1e1fb400821b9f241038a67fac9b334b6340c7" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.932036 4959 scope.go:117] "RemoveContainer" containerID="63f045b52f496803aa5e9e76437ab3594521f8d8772b5b88332ebc178da54515" Oct 07 13:45:55 crc kubenswrapper[4959]: E1007 13:45:55.932403 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5sfv5_openshift-ovn-kubernetes(b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.945253 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.945304 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.945322 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.945344 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.945362 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:55Z","lastTransitionTime":"2025-10-07T13:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.964614 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:55Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:55 crc kubenswrapper[4959]: I1007 13:45:55.984499 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:55Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.000838 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:55Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.016958 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:56Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.035813 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:56Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.048023 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.048085 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.048130 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.048156 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.048173 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:56Z","lastTransitionTime":"2025-10-07T13:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.052948 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:56Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.073402 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:56Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.090786 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:56Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.111903 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:56Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.130766 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:56Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.151418 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.151483 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.151504 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.151530 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.151548 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:56Z","lastTransitionTime":"2025-10-07T13:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.153940 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1cbef019f0640b5d51b398d1fc7166c1e253fd33f78f0e50f9754bb1187a6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:56Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.172576 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:56Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.189461 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:56Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.207076 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:56Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.237719 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63f045b52f496803aa5e9e76437ab3594521f8d8772b5b88332ebc178da54515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee769877d15b988363ec0512da1e1fb400821b9f241038a67fac9b334b6340c7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:45:54Z\\\",\\\"message\\\":\\\" 6251 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1007 13:45:54.475870 6251 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1007 13:45:54.475918 6251 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1007 13:45:54.475944 6251 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1007 13:45:54.475980 6251 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1007 13:45:54.475989 6251 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1007 13:45:54.476011 6251 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 13:45:54.476021 6251 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1007 13:45:54.476079 6251 factory.go:656] Stopping watch factory\\\\nI1007 13:45:54.476134 6251 ovnkube.go:599] Stopped ovnkube\\\\nI1007 13:45:54.476179 6251 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1007 13:45:54.476195 6251 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1007 13:45:54.476209 6251 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1007 13:45:54.476223 6251 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1007 13:45:54.476238 6251 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1007 13:45:54.476252 6251 handler.go:208] Removed *v1.Node event handler 2\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63f045b52f496803aa5e9e76437ab3594521f8d8772b5b88332ebc178da54515\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:45:55Z\\\",\\\"message\\\":\\\"o:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1007 13:45:55.648874 6373 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1007 13:45:55.648891 6373 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1007 13:45:55.648894 6373 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1007 13:45:55.648900 6373 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI1007 13:45:55.648902 6373 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1007 13:45:55.648907 6373 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nF1007 13:45:55.648882 6373 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal e\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:56Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.254661 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.254726 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.254743 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.254795 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.254813 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:56Z","lastTransitionTime":"2025-10-07T13:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.357887 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.357968 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.357986 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.358013 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.358034 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:56Z","lastTransitionTime":"2025-10-07T13:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.460415 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.460472 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.460493 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.460518 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.460536 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:56Z","lastTransitionTime":"2025-10-07T13:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.563437 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.563512 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.563539 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.563575 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.563603 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:56Z","lastTransitionTime":"2025-10-07T13:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.652882 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.652901 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:45:56 crc kubenswrapper[4959]: E1007 13:45:56.653170 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:45:56 crc kubenswrapper[4959]: E1007 13:45:56.653297 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.666233 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.666283 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.666295 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.666312 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.666325 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:56Z","lastTransitionTime":"2025-10-07T13:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.769367 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.769418 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.769432 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.769448 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.769474 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:56Z","lastTransitionTime":"2025-10-07T13:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.872614 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.872672 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.872693 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.872724 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.872742 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:56Z","lastTransitionTime":"2025-10-07T13:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.936013 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5sfv5_b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b/ovnkube-controller/1.log" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.939020 4959 scope.go:117] "RemoveContainer" containerID="63f045b52f496803aa5e9e76437ab3594521f8d8772b5b88332ebc178da54515" Oct 07 13:45:56 crc kubenswrapper[4959]: E1007 13:45:56.939226 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5sfv5_openshift-ovn-kubernetes(b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.952544 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:56Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.968855 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:56Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.975640 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.975676 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.975687 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.975702 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.975712 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:56Z","lastTransitionTime":"2025-10-07T13:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:56 crc kubenswrapper[4959]: I1007 13:45:56.988778 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:56Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.011949 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63f045b52f496803aa5e9e76437ab3594521f8d8772b5b88332ebc178da54515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63f045b52f496803aa5e9e76437ab3594521f8d8772b5b88332ebc178da54515\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:45:55Z\\\",\\\"message\\\":\\\"o:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1007 13:45:55.648874 6373 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1007 13:45:55.648891 6373 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1007 13:45:55.648894 6373 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1007 13:45:55.648900 6373 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI1007 13:45:55.648902 6373 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1007 13:45:55.648907 6373 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nF1007 13:45:55.648882 6373 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal e\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5sfv5_openshift-ovn-kubernetes(b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:57Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.025388 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:57Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.040611 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:57Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.051413 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:57Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.062194 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:57Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.078270 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.078324 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.078339 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.078359 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.078372 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:57Z","lastTransitionTime":"2025-10-07T13:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.093203 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:57Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.115712 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:57Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.134850 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:57Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.155464 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:57Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.173649 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:57Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.181468 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.181511 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.181523 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.181541 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.181553 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:57Z","lastTransitionTime":"2025-10-07T13:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.194961 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1cbef019f0640b5d51b398d1fc7166c1e253fd33f78f0e50f9754bb1187a6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:57Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.207897 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:57Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.284325 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.284363 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.284372 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.284389 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.284400 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:57Z","lastTransitionTime":"2025-10-07T13:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.387245 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.387306 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.387322 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.387398 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.387418 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:57Z","lastTransitionTime":"2025-10-07T13:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.489576 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.489638 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.489654 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.489670 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.489685 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:57Z","lastTransitionTime":"2025-10-07T13:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.592751 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.592790 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.592799 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.592812 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.592821 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:57Z","lastTransitionTime":"2025-10-07T13:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.652935 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:45:57 crc kubenswrapper[4959]: E1007 13:45:57.653078 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.695230 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.695280 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.695292 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.695311 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.695323 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:57Z","lastTransitionTime":"2025-10-07T13:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.763080 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs"] Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.764260 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.767499 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.767692 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.789739 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:57Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.797691 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.797740 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.797757 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.797785 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.797803 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:57Z","lastTransitionTime":"2025-10-07T13:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.817333 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1cbef019f0640b5d51b398d1fc7166c1e253fd33f78f0e50f9754bb1187a6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:57Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.838701 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:57Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.860216 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:57Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.879836 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:57Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.899983 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:57Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.901460 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.901513 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.901531 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.901555 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.901572 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:57Z","lastTransitionTime":"2025-10-07T13:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.932446 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-4xpjs\" (UID: \"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.932547 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-4xpjs\" (UID: \"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.932568 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-4xpjs\" (UID: \"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.932629 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmsrh\" (UniqueName: \"kubernetes.io/projected/9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f-kube-api-access-kmsrh\") pod \"ovnkube-control-plane-749d76644c-4xpjs\" (UID: \"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.940939 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63f045b52f496803aa5e9e76437ab3594521f8d8772b5b88332ebc178da54515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63f045b52f496803aa5e9e76437ab3594521f8d8772b5b88332ebc178da54515\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:45:55Z\\\",\\\"message\\\":\\\"o:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1007 13:45:55.648874 6373 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1007 13:45:55.648891 6373 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1007 13:45:55.648894 6373 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1007 13:45:55.648900 6373 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI1007 13:45:55.648902 6373 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1007 13:45:55.648907 6373 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nF1007 13:45:55.648882 6373 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal e\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5sfv5_openshift-ovn-kubernetes(b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:57Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.963167 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:57Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:57 crc kubenswrapper[4959]: I1007 13:45:57.981049 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:57Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.002759 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:57Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.003922 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.003991 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.004014 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.004040 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.004058 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:58Z","lastTransitionTime":"2025-10-07T13:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.020491 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.033649 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmsrh\" (UniqueName: \"kubernetes.io/projected/9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f-kube-api-access-kmsrh\") pod \"ovnkube-control-plane-749d76644c-4xpjs\" (UID: \"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.033712 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-4xpjs\" (UID: \"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.033779 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-4xpjs\" (UID: \"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.033813 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-4xpjs\" (UID: \"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.035144 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-4xpjs\" (UID: \"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.035280 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-4xpjs\" (UID: \"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.043856 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xpjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.045979 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-4xpjs\" (UID: \"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.055948 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmsrh\" (UniqueName: \"kubernetes.io/projected/9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f-kube-api-access-kmsrh\") pod \"ovnkube-control-plane-749d76644c-4xpjs\" (UID: \"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.074931 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.085973 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.095826 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.107194 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.107564 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.107577 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.107595 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.107607 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:58Z","lastTransitionTime":"2025-10-07T13:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.119378 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.139506 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.213470 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.213505 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.213516 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.213530 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.213540 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:58Z","lastTransitionTime":"2025-10-07T13:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.315861 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.315898 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.315906 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.315918 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.315926 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:58Z","lastTransitionTime":"2025-10-07T13:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.417814 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.417855 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.417866 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.417882 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.417895 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:58Z","lastTransitionTime":"2025-10-07T13:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.520651 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.520704 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.520721 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.520739 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.520754 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:58Z","lastTransitionTime":"2025-10-07T13:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.583127 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-zdps5"] Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.583652 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:45:58 crc kubenswrapper[4959]: E1007 13:45:58.583725 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.598240 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.611726 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.622972 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.623022 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.623037 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.623056 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.623072 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:58Z","lastTransitionTime":"2025-10-07T13:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.625626 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.639342 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.644548 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fnzz\" (UniqueName: \"kubernetes.io/projected/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-kube-api-access-9fnzz\") pod \"network-metrics-daemon-zdps5\" (UID: \"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\") " pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.644602 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-metrics-certs\") pod \"network-metrics-daemon-zdps5\" (UID: \"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\") " pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.652611 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.652629 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:45:58 crc kubenswrapper[4959]: E1007 13:45:58.652726 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:45:58 crc kubenswrapper[4959]: E1007 13:45:58.652847 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.656000 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1cbef019f0640b5d51b398d1fc7166c1e253fd33f78f0e50f9754bb1187a6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.667049 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.681651 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.696671 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.709592 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.725508 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.725547 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.725561 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.725578 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.725592 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:58Z","lastTransitionTime":"2025-10-07T13:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.726701 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63f045b52f496803aa5e9e76437ab3594521f8d8772b5b88332ebc178da54515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63f045b52f496803aa5e9e76437ab3594521f8d8772b5b88332ebc178da54515\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:45:55Z\\\",\\\"message\\\":\\\"o:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1007 13:45:55.648874 6373 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1007 13:45:55.648891 6373 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1007 13:45:55.648894 6373 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1007 13:45:55.648900 6373 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI1007 13:45:55.648902 6373 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1007 13:45:55.648907 6373 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nF1007 13:45:55.648882 6373 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal e\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5sfv5_openshift-ovn-kubernetes(b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.745218 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-metrics-certs\") pod \"network-metrics-daemon-zdps5\" (UID: \"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\") " pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:45:58 crc kubenswrapper[4959]: E1007 13:45:58.745341 4959 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 13:45:58 crc kubenswrapper[4959]: E1007 13:45:58.745402 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-metrics-certs podName:50d4ac5c-0a16-4530-8500-9975eb2b6b5a nodeName:}" failed. No retries permitted until 2025-10-07 13:45:59.245388141 +0000 UTC m=+41.328792466 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-metrics-certs") pod "network-metrics-daemon-zdps5" (UID: "50d4ac5c-0a16-4530-8500-9975eb2b6b5a") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.745395 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fnzz\" (UniqueName: \"kubernetes.io/projected/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-kube-api-access-9fnzz\") pod \"network-metrics-daemon-zdps5\" (UID: \"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\") " pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.748169 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.759559 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.762075 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fnzz\" (UniqueName: \"kubernetes.io/projected/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-kube-api-access-9fnzz\") pod \"network-metrics-daemon-zdps5\" (UID: \"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\") " pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.772071 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.781934 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.791580 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.807123 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xpjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.817515 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zdps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zdps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.829261 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.829295 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.829303 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.829319 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.829328 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:58Z","lastTransitionTime":"2025-10-07T13:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.831405 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.844919 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1cbef019f0640b5d51b398d1fc7166c1e253fd33f78f0e50f9754bb1187a6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.854331 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.869356 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.881281 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.891110 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.911297 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63f045b52f496803aa5e9e76437ab3594521f8d8772b5b88332ebc178da54515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63f045b52f496803aa5e9e76437ab3594521f8d8772b5b88332ebc178da54515\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:45:55Z\\\",\\\"message\\\":\\\"o:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1007 13:45:55.648874 6373 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1007 13:45:55.648891 6373 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1007 13:45:55.648894 6373 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1007 13:45:55.648900 6373 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI1007 13:45:55.648902 6373 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1007 13:45:55.648907 6373 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nF1007 13:45:55.648882 6373 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal e\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5sfv5_openshift-ovn-kubernetes(b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.923163 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.932380 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.932450 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.932467 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.932490 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.932508 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:58Z","lastTransitionTime":"2025-10-07T13:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.933856 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.947408 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.948322 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" event={"ID":"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f","Type":"ContainerStarted","Data":"1454006ab0b9f83bd85e57a5e7389c81aa81ea5fb4d7c86a6315839de5c37812"} Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.948497 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" event={"ID":"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f","Type":"ContainerStarted","Data":"34ab500b1b215518cbfc4602ee0671c6f3f68464467aa392ea5bc38ffb24bd8b"} Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.948681 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" event={"ID":"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f","Type":"ContainerStarted","Data":"48917683625a1db835fd9352be99e3f1fb1ebc29f2f1cb79d00f34204953c3d2"} Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.959055 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.969852 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xpjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:58 crc kubenswrapper[4959]: I1007 13:45:58.981013 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zdps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zdps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.011954 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:59Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.024524 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:59Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.035697 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.035730 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.035739 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.035752 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.035765 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:59Z","lastTransitionTime":"2025-10-07T13:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.038668 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:59Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.051207 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:59Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.065414 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:59Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.081057 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:59Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.095084 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:59Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.122985 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63f045b52f496803aa5e9e76437ab3594521f8d8772b5b88332ebc178da54515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63f045b52f496803aa5e9e76437ab3594521f8d8772b5b88332ebc178da54515\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:45:55Z\\\",\\\"message\\\":\\\"o:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1007 13:45:55.648874 6373 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1007 13:45:55.648891 6373 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1007 13:45:55.648894 6373 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1007 13:45:55.648900 6373 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI1007 13:45:55.648902 6373 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1007 13:45:55.648907 6373 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nF1007 13:45:55.648882 6373 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal e\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5sfv5_openshift-ovn-kubernetes(b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:59Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.138725 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.138786 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.138804 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.138828 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.138846 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:59Z","lastTransitionTime":"2025-10-07T13:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.140046 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ab500b1b215518cbfc4602ee0671c6f3f68464467aa392ea5bc38ffb24bd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1454006ab0b9f83bd85e57a5e7389c81aa81ea5fb4d7c86a6315839de5c37812\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xpjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:59Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.154685 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zdps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zdps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:59Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.180026 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:59Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.193281 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:59Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.214513 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:59Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.226083 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:59Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.239627 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:59Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.241381 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.241437 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.241458 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.241486 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.241509 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:59Z","lastTransitionTime":"2025-10-07T13:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.249080 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-metrics-certs\") pod \"network-metrics-daemon-zdps5\" (UID: \"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\") " pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:45:59 crc kubenswrapper[4959]: E1007 13:45:59.249195 4959 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 13:45:59 crc kubenswrapper[4959]: E1007 13:45:59.249247 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-metrics-certs podName:50d4ac5c-0a16-4530-8500-9975eb2b6b5a nodeName:}" failed. No retries permitted until 2025-10-07 13:46:00.249233627 +0000 UTC m=+42.332637952 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-metrics-certs") pod "network-metrics-daemon-zdps5" (UID: "50d4ac5c-0a16-4530-8500-9975eb2b6b5a") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.259419 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1cbef019f0640b5d51b398d1fc7166c1e253fd33f78f0e50f9754bb1187a6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:59Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.272219 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:59Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.290909 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:59Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.304051 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:59Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.318897 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:59Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.333673 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:45:59Z is after 2025-08-24T17:21:41Z" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.344526 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.344643 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.344715 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.344796 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.344873 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:59Z","lastTransitionTime":"2025-10-07T13:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.447819 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.447849 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.447857 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.447870 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.447880 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:59Z","lastTransitionTime":"2025-10-07T13:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.550547 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.550614 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.550638 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.550669 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.550691 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:59Z","lastTransitionTime":"2025-10-07T13:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.652603 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:45:59 crc kubenswrapper[4959]: E1007 13:45:59.652826 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.652991 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.653044 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.653063 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.653085 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.653135 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:59Z","lastTransitionTime":"2025-10-07T13:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.756813 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.756891 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.756911 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.756939 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.756960 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:59Z","lastTransitionTime":"2025-10-07T13:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.859662 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.859725 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.859744 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.859769 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.859787 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:59Z","lastTransitionTime":"2025-10-07T13:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.962318 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.962374 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.962390 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.962415 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:45:59 crc kubenswrapper[4959]: I1007 13:45:59.962432 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:45:59Z","lastTransitionTime":"2025-10-07T13:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.065642 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.065950 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.066158 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.066338 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.066469 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:00Z","lastTransitionTime":"2025-10-07T13:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.169512 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.169577 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.169600 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.169627 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.169648 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:00Z","lastTransitionTime":"2025-10-07T13:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.260480 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-metrics-certs\") pod \"network-metrics-daemon-zdps5\" (UID: \"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\") " pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:00 crc kubenswrapper[4959]: E1007 13:46:00.260758 4959 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 13:46:00 crc kubenswrapper[4959]: E1007 13:46:00.261260 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-metrics-certs podName:50d4ac5c-0a16-4530-8500-9975eb2b6b5a nodeName:}" failed. No retries permitted until 2025-10-07 13:46:02.261231567 +0000 UTC m=+44.344635932 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-metrics-certs") pod "network-metrics-daemon-zdps5" (UID: "50d4ac5c-0a16-4530-8500-9975eb2b6b5a") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.272729 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.272792 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.272813 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.272839 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.272857 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:00Z","lastTransitionTime":"2025-10-07T13:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.374785 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.374817 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.374825 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.374836 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.374844 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:00Z","lastTransitionTime":"2025-10-07T13:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.478157 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.478207 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.478218 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.478238 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.478249 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:00Z","lastTransitionTime":"2025-10-07T13:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.581371 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.581412 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.581425 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.581442 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.581455 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:00Z","lastTransitionTime":"2025-10-07T13:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.653239 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.653319 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:00 crc kubenswrapper[4959]: E1007 13:46:00.653345 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:00 crc kubenswrapper[4959]: E1007 13:46:00.653453 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.653503 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:00 crc kubenswrapper[4959]: E1007 13:46:00.653685 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.684241 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.684284 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.684296 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.684315 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.684327 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:00Z","lastTransitionTime":"2025-10-07T13:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.786978 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.787013 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.787024 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.787040 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.787051 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:00Z","lastTransitionTime":"2025-10-07T13:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.889807 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.889845 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.889856 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.889871 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.889883 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:00Z","lastTransitionTime":"2025-10-07T13:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.993304 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.993367 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.993384 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.993409 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:00 crc kubenswrapper[4959]: I1007 13:46:00.993427 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:00Z","lastTransitionTime":"2025-10-07T13:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.096146 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.096198 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.096215 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.096237 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.096254 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:01Z","lastTransitionTime":"2025-10-07T13:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.203396 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.203462 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.203487 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.203517 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.203554 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:01Z","lastTransitionTime":"2025-10-07T13:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.306945 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.307001 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.307018 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.307045 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.307064 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:01Z","lastTransitionTime":"2025-10-07T13:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.410680 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.410744 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.410762 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.410785 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.410803 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:01Z","lastTransitionTime":"2025-10-07T13:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.514342 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.514401 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.514417 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.514441 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.514457 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:01Z","lastTransitionTime":"2025-10-07T13:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.618397 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.618440 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.618452 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.618476 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.618493 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:01Z","lastTransitionTime":"2025-10-07T13:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.653001 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:01 crc kubenswrapper[4959]: E1007 13:46:01.653165 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.721207 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.721245 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.721256 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.721272 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.721283 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:01Z","lastTransitionTime":"2025-10-07T13:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.823746 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.823824 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.823848 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.823879 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.823907 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:01Z","lastTransitionTime":"2025-10-07T13:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.926544 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.926616 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.926636 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.926665 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:01 crc kubenswrapper[4959]: I1007 13:46:01.926687 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:01Z","lastTransitionTime":"2025-10-07T13:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.030404 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.030478 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.030497 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.030525 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.030543 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:02Z","lastTransitionTime":"2025-10-07T13:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.133955 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.134001 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.134018 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.134041 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.134058 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:02Z","lastTransitionTime":"2025-10-07T13:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.237675 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.237736 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.237758 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.237785 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.237820 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:02Z","lastTransitionTime":"2025-10-07T13:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.283766 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-metrics-certs\") pod \"network-metrics-daemon-zdps5\" (UID: \"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\") " pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:02 crc kubenswrapper[4959]: E1007 13:46:02.283968 4959 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 13:46:02 crc kubenswrapper[4959]: E1007 13:46:02.284088 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-metrics-certs podName:50d4ac5c-0a16-4530-8500-9975eb2b6b5a nodeName:}" failed. No retries permitted until 2025-10-07 13:46:06.284055306 +0000 UTC m=+48.367459671 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-metrics-certs") pod "network-metrics-daemon-zdps5" (UID: "50d4ac5c-0a16-4530-8500-9975eb2b6b5a") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.340236 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.340281 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.340290 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.340305 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.340321 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:02Z","lastTransitionTime":"2025-10-07T13:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.443429 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.443504 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.443529 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.443556 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.443577 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:02Z","lastTransitionTime":"2025-10-07T13:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.545928 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.545988 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.546004 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.546029 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.546046 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:02Z","lastTransitionTime":"2025-10-07T13:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.648820 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.648887 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.648910 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.648941 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.648965 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:02Z","lastTransitionTime":"2025-10-07T13:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.653200 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.653300 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.653225 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:02 crc kubenswrapper[4959]: E1007 13:46:02.653408 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:02 crc kubenswrapper[4959]: E1007 13:46:02.653556 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:02 crc kubenswrapper[4959]: E1007 13:46:02.653735 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.752340 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.752411 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.752432 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.752462 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.752483 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:02Z","lastTransitionTime":"2025-10-07T13:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.856160 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.856223 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.856243 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.856269 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.856288 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:02Z","lastTransitionTime":"2025-10-07T13:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.959824 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.959882 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.959899 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.959923 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:02 crc kubenswrapper[4959]: I1007 13:46:02.959942 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:02Z","lastTransitionTime":"2025-10-07T13:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.063624 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.063678 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.063696 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.063718 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.063736 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:03Z","lastTransitionTime":"2025-10-07T13:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.167270 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.167346 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.167371 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.167398 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.167419 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:03Z","lastTransitionTime":"2025-10-07T13:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.269979 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.270027 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.270049 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.270068 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.270084 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:03Z","lastTransitionTime":"2025-10-07T13:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.373016 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.373078 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.373094 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.373137 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.373152 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:03Z","lastTransitionTime":"2025-10-07T13:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.476359 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.476446 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.476476 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.476507 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.476530 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:03Z","lastTransitionTime":"2025-10-07T13:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.579571 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.579662 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.579701 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.579735 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.579759 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:03Z","lastTransitionTime":"2025-10-07T13:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.653478 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:03 crc kubenswrapper[4959]: E1007 13:46:03.653686 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.682518 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.682578 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.682598 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.682640 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.682698 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:03Z","lastTransitionTime":"2025-10-07T13:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.786631 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.786692 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.786743 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.786768 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.786787 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:03Z","lastTransitionTime":"2025-10-07T13:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.890055 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.890151 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.890170 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.890188 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.890226 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:03Z","lastTransitionTime":"2025-10-07T13:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.993825 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.993887 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.993905 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.993932 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:03 crc kubenswrapper[4959]: I1007 13:46:03.993951 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:03Z","lastTransitionTime":"2025-10-07T13:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.096494 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.096546 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.096563 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.096588 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.096606 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:04Z","lastTransitionTime":"2025-10-07T13:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.200604 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.200648 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.200659 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.200675 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.200684 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:04Z","lastTransitionTime":"2025-10-07T13:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.304596 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.304650 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.304665 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.304685 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.304698 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:04Z","lastTransitionTime":"2025-10-07T13:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.408154 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.408231 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.408247 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.408270 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.408283 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:04Z","lastTransitionTime":"2025-10-07T13:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.511337 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.511396 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.511413 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.511436 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.511453 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:04Z","lastTransitionTime":"2025-10-07T13:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.615313 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.615372 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.615389 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.615415 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.615436 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:04Z","lastTransitionTime":"2025-10-07T13:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.653483 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.653511 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:04 crc kubenswrapper[4959]: E1007 13:46:04.653597 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:04 crc kubenswrapper[4959]: E1007 13:46:04.653775 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.653845 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:04 crc kubenswrapper[4959]: E1007 13:46:04.653900 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.718010 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.718075 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.718092 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.718140 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.718159 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:04Z","lastTransitionTime":"2025-10-07T13:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.821370 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.821427 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.821444 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.821467 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.821485 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:04Z","lastTransitionTime":"2025-10-07T13:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.847567 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.847624 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.847642 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.847673 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.847691 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:04Z","lastTransitionTime":"2025-10-07T13:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:04 crc kubenswrapper[4959]: E1007 13:46:04.862190 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:04Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.867554 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.867623 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.867640 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.867664 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.867687 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:04Z","lastTransitionTime":"2025-10-07T13:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:04 crc kubenswrapper[4959]: E1007 13:46:04.882763 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:04Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.887765 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.887840 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.887872 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.887901 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.887923 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:04Z","lastTransitionTime":"2025-10-07T13:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:04 crc kubenswrapper[4959]: E1007 13:46:04.901954 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:04Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.906143 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.906221 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.906245 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.906274 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.906297 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:04Z","lastTransitionTime":"2025-10-07T13:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:04 crc kubenswrapper[4959]: E1007 13:46:04.921883 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:04Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.925981 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.926050 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.926075 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.926138 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.926167 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:04Z","lastTransitionTime":"2025-10-07T13:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:04 crc kubenswrapper[4959]: E1007 13:46:04.937231 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:04Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:04 crc kubenswrapper[4959]: E1007 13:46:04.937392 4959 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.939502 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.939535 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.939546 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.939560 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:04 crc kubenswrapper[4959]: I1007 13:46:04.939573 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:04Z","lastTransitionTime":"2025-10-07T13:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.042153 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.042205 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.042215 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.042230 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.042239 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:05Z","lastTransitionTime":"2025-10-07T13:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.145588 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.145659 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.145678 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.145708 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.145789 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:05Z","lastTransitionTime":"2025-10-07T13:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.248931 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.248971 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.248982 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.249016 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.249027 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:05Z","lastTransitionTime":"2025-10-07T13:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.352038 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.352396 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.352407 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.352424 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.352437 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:05Z","lastTransitionTime":"2025-10-07T13:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.455742 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.455825 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.455849 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.455879 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.455901 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:05Z","lastTransitionTime":"2025-10-07T13:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.559466 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.559565 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.559594 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.559629 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.559679 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:05Z","lastTransitionTime":"2025-10-07T13:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.652816 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:05 crc kubenswrapper[4959]: E1007 13:46:05.653074 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.663838 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.663896 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.663914 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.663944 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.663964 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:05Z","lastTransitionTime":"2025-10-07T13:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.767963 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.768038 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.768060 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.768089 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.768154 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:05Z","lastTransitionTime":"2025-10-07T13:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.872201 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.872255 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.872271 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.872292 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.872307 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:05Z","lastTransitionTime":"2025-10-07T13:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.975603 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.975662 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.975680 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.975710 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:05 crc kubenswrapper[4959]: I1007 13:46:05.975732 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:05Z","lastTransitionTime":"2025-10-07T13:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.080441 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.080525 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.080543 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.080571 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.080593 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:06Z","lastTransitionTime":"2025-10-07T13:46:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.183619 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.183682 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.183729 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.183761 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.183783 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:06Z","lastTransitionTime":"2025-10-07T13:46:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.287019 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.287091 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.287136 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.287226 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.287300 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:06Z","lastTransitionTime":"2025-10-07T13:46:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.329976 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-metrics-certs\") pod \"network-metrics-daemon-zdps5\" (UID: \"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\") " pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:06 crc kubenswrapper[4959]: E1007 13:46:06.330342 4959 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 13:46:06 crc kubenswrapper[4959]: E1007 13:46:06.330479 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-metrics-certs podName:50d4ac5c-0a16-4530-8500-9975eb2b6b5a nodeName:}" failed. No retries permitted until 2025-10-07 13:46:14.330446624 +0000 UTC m=+56.413850989 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-metrics-certs") pod "network-metrics-daemon-zdps5" (UID: "50d4ac5c-0a16-4530-8500-9975eb2b6b5a") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.389893 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.389927 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.389938 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.389955 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.389965 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:06Z","lastTransitionTime":"2025-10-07T13:46:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.492787 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.492858 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.492883 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.492914 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.492932 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:06Z","lastTransitionTime":"2025-10-07T13:46:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.596453 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.596497 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.596505 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.596521 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.596532 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:06Z","lastTransitionTime":"2025-10-07T13:46:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.652604 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.652738 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:06 crc kubenswrapper[4959]: E1007 13:46:06.652874 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.652925 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:06 crc kubenswrapper[4959]: E1007 13:46:06.653065 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:06 crc kubenswrapper[4959]: E1007 13:46:06.653266 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.700859 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.700914 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.700932 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.700962 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.700982 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:06Z","lastTransitionTime":"2025-10-07T13:46:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.803571 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.803609 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.803618 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.803633 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.803644 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:06Z","lastTransitionTime":"2025-10-07T13:46:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.906189 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.906463 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.906482 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.906512 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:06 crc kubenswrapper[4959]: I1007 13:46:06.906534 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:06Z","lastTransitionTime":"2025-10-07T13:46:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.009855 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.009905 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.009921 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.009945 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.009962 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:07Z","lastTransitionTime":"2025-10-07T13:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.114531 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.114587 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.114604 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.114627 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.114648 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:07Z","lastTransitionTime":"2025-10-07T13:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.217800 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.217861 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.217877 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.217901 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.217919 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:07Z","lastTransitionTime":"2025-10-07T13:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.321339 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.321381 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.321395 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.321413 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.321424 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:07Z","lastTransitionTime":"2025-10-07T13:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.423857 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.423894 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.423903 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.423917 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.423925 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:07Z","lastTransitionTime":"2025-10-07T13:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.527017 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.527059 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.527071 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.527088 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.527119 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:07Z","lastTransitionTime":"2025-10-07T13:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.630495 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.630545 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.630556 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.630574 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.630589 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:07Z","lastTransitionTime":"2025-10-07T13:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.653395 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:07 crc kubenswrapper[4959]: E1007 13:46:07.653537 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.733740 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.733795 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.733813 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.733834 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.733848 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:07Z","lastTransitionTime":"2025-10-07T13:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.836459 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.836524 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.836542 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.836566 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.836582 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:07Z","lastTransitionTime":"2025-10-07T13:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.939443 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.939503 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.939519 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.939545 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:07 crc kubenswrapper[4959]: I1007 13:46:07.939574 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:07Z","lastTransitionTime":"2025-10-07T13:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.041623 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.041666 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.041675 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.041689 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.041699 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:08Z","lastTransitionTime":"2025-10-07T13:46:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.144059 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.144090 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.144118 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.144131 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.144139 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:08Z","lastTransitionTime":"2025-10-07T13:46:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.245885 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.245950 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.245968 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.245993 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.246011 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:08Z","lastTransitionTime":"2025-10-07T13:46:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.348812 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.348859 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.348871 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.348890 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.348904 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:08Z","lastTransitionTime":"2025-10-07T13:46:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.394895 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.405991 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.410225 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.427917 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.445769 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.450775 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.450931 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.450962 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.450996 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.451019 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:08Z","lastTransitionTime":"2025-10-07T13:46:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.464434 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.484609 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.506388 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1cbef019f0640b5d51b398d1fc7166c1e253fd33f78f0e50f9754bb1187a6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.518700 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.530136 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.542514 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.553499 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.553548 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.553566 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.553590 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.553605 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:08Z","lastTransitionTime":"2025-10-07T13:46:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.561020 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63f045b52f496803aa5e9e76437ab3594521f8d8772b5b88332ebc178da54515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63f045b52f496803aa5e9e76437ab3594521f8d8772b5b88332ebc178da54515\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:45:55Z\\\",\\\"message\\\":\\\"o:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1007 13:45:55.648874 6373 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1007 13:45:55.648891 6373 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1007 13:45:55.648894 6373 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1007 13:45:55.648900 6373 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI1007 13:45:55.648902 6373 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1007 13:45:55.648907 6373 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nF1007 13:45:55.648882 6373 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal e\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5sfv5_openshift-ovn-kubernetes(b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.572624 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zdps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zdps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.591919 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.604375 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.616167 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.625669 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.635590 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.645542 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ab500b1b215518cbfc4602ee0671c6f3f68464467aa392ea5bc38ffb24bd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1454006ab0b9f83bd85e57a5e7389c81aa81ea5fb4d7c86a6315839de5c37812\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xpjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.652850 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.652927 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.652865 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:08 crc kubenswrapper[4959]: E1007 13:46:08.652999 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:08 crc kubenswrapper[4959]: E1007 13:46:08.653075 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:08 crc kubenswrapper[4959]: E1007 13:46:08.653531 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.653915 4959 scope.go:117] "RemoveContainer" containerID="63f045b52f496803aa5e9e76437ab3594521f8d8772b5b88332ebc178da54515" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.657010 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.657148 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.657236 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.657361 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.657433 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:08Z","lastTransitionTime":"2025-10-07T13:46:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.665658 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.678985 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.691766 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.707246 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1cbef019f0640b5d51b398d1fc7166c1e253fd33f78f0e50f9754bb1187a6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.719404 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.743702 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.760067 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.760128 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.760144 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.760167 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.760183 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:08Z","lastTransitionTime":"2025-10-07T13:46:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.763001 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.785022 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.809685 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63f045b52f496803aa5e9e76437ab3594521f8d8772b5b88332ebc178da54515\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63f045b52f496803aa5e9e76437ab3594521f8d8772b5b88332ebc178da54515\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:45:55Z\\\",\\\"message\\\":\\\"o:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1007 13:45:55.648874 6373 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1007 13:45:55.648891 6373 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1007 13:45:55.648894 6373 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1007 13:45:55.648900 6373 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI1007 13:45:55.648902 6373 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1007 13:45:55.648907 6373 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nF1007 13:45:55.648882 6373 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal e\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5sfv5_openshift-ovn-kubernetes(b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.825318 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e325099f-e0ff-433d-bef7-6909c5c86bea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecb7da009b7ed9e5c0412125be08cf486bf4f0e74480f2843877b0816944fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6dab7bc77a409dca97cffd302d0086021bf51146e8ec69a53cbddb87a4026aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b500df391c3b1ef8f9689a04e58c19e603667507a68fb8d65ea768690425085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.843664 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.863070 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.863696 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.863721 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.863734 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.863752 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.863764 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:08Z","lastTransitionTime":"2025-10-07T13:46:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.879228 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.895721 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.909474 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ab500b1b215518cbfc4602ee0671c6f3f68464467aa392ea5bc38ffb24bd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1454006ab0b9f83bd85e57a5e7389c81aa81ea5fb4d7c86a6315839de5c37812\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xpjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.923652 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zdps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zdps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.943546 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.958619 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.966807 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.966850 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.966862 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.966882 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.966896 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:08Z","lastTransitionTime":"2025-10-07T13:46:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.989494 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5sfv5_b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b/ovnkube-controller/1.log" Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.992879 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" event={"ID":"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b","Type":"ContainerStarted","Data":"261a778f70ae84c72b7224921a7f335d85bff09dc9ab003434f344cec3f78161"} Oct 07 13:46:08 crc kubenswrapper[4959]: I1007 13:46:08.993703 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.006523 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:09Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.018491 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:09Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.031468 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:09Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.050040 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:09Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.069845 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.069888 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.069900 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.069922 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.069937 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:09Z","lastTransitionTime":"2025-10-07T13:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.078145 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1cbef019f0640b5d51b398d1fc7166c1e253fd33f78f0e50f9754bb1187a6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:09Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.091844 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:09Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.105625 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:09Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.174168 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:09Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.175666 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.175698 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.175707 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.175721 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.175751 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:09Z","lastTransitionTime":"2025-10-07T13:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.187924 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:09Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.218993 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://261a778f70ae84c72b7224921a7f335d85bff09dc9ab003434f344cec3f78161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63f045b52f496803aa5e9e76437ab3594521f8d8772b5b88332ebc178da54515\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:45:55Z\\\",\\\"message\\\":\\\"o:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1007 13:45:55.648874 6373 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1007 13:45:55.648891 6373 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1007 13:45:55.648894 6373 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1007 13:45:55.648900 6373 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI1007 13:45:55.648902 6373 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1007 13:45:55.648907 6373 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nF1007 13:45:55.648882 6373 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal e\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:46:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:09Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.235159 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e325099f-e0ff-433d-bef7-6909c5c86bea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecb7da009b7ed9e5c0412125be08cf486bf4f0e74480f2843877b0816944fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6dab7bc77a409dca97cffd302d0086021bf51146e8ec69a53cbddb87a4026aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b500df391c3b1ef8f9689a04e58c19e603667507a68fb8d65ea768690425085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:09Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.252579 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:09Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.270198 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:09Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.280304 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.280343 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.280353 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.280370 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.280383 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:09Z","lastTransitionTime":"2025-10-07T13:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.290584 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:09Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.305273 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:09Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.316494 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ab500b1b215518cbfc4602ee0671c6f3f68464467aa392ea5bc38ffb24bd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1454006ab0b9f83bd85e57a5e7389c81aa81ea5fb4d7c86a6315839de5c37812\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xpjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:09Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.329756 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zdps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zdps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:09Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.348294 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:09Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.381973 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.382019 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.382028 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.382045 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.382056 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:09Z","lastTransitionTime":"2025-10-07T13:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.484228 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.484276 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.484286 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.484305 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.484316 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:09Z","lastTransitionTime":"2025-10-07T13:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.587072 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.587142 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.587160 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.587182 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.587197 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:09Z","lastTransitionTime":"2025-10-07T13:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.652794 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:09 crc kubenswrapper[4959]: E1007 13:46:09.652937 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.689630 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.689678 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.689696 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.689718 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.689735 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:09Z","lastTransitionTime":"2025-10-07T13:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.792651 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.792689 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.792699 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.792714 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.792725 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:09Z","lastTransitionTime":"2025-10-07T13:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.896149 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.896237 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.896262 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.896295 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.896323 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:09Z","lastTransitionTime":"2025-10-07T13:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.998430 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.998484 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.998502 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.998524 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.998543 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:09Z","lastTransitionTime":"2025-10-07T13:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:09 crc kubenswrapper[4959]: I1007 13:46:09.999929 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5sfv5_b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b/ovnkube-controller/2.log" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.000986 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5sfv5_b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b/ovnkube-controller/1.log" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.004889 4959 generic.go:334] "Generic (PLEG): container finished" podID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerID="261a778f70ae84c72b7224921a7f335d85bff09dc9ab003434f344cec3f78161" exitCode=1 Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.004954 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" event={"ID":"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b","Type":"ContainerDied","Data":"261a778f70ae84c72b7224921a7f335d85bff09dc9ab003434f344cec3f78161"} Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.005016 4959 scope.go:117] "RemoveContainer" containerID="63f045b52f496803aa5e9e76437ab3594521f8d8772b5b88332ebc178da54515" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.006139 4959 scope.go:117] "RemoveContainer" containerID="261a778f70ae84c72b7224921a7f335d85bff09dc9ab003434f344cec3f78161" Oct 07 13:46:10 crc kubenswrapper[4959]: E1007 13:46:10.006391 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5sfv5_openshift-ovn-kubernetes(b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.028469 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e325099f-e0ff-433d-bef7-6909c5c86bea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecb7da009b7ed9e5c0412125be08cf486bf4f0e74480f2843877b0816944fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6dab7bc77a409dca97cffd302d0086021bf51146e8ec69a53cbddb87a4026aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b500df391c3b1ef8f9689a04e58c19e603667507a68fb8d65ea768690425085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:10Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.049349 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:10Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.063718 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:10Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.076719 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:10Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.101138 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.101205 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.101222 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.101247 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.101266 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:10Z","lastTransitionTime":"2025-10-07T13:46:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.106835 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://261a778f70ae84c72b7224921a7f335d85bff09dc9ab003434f344cec3f78161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63f045b52f496803aa5e9e76437ab3594521f8d8772b5b88332ebc178da54515\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:45:55Z\\\",\\\"message\\\":\\\"o:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1007 13:45:55.648874 6373 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1007 13:45:55.648891 6373 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1007 13:45:55.648894 6373 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1007 13:45:55.648900 6373 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI1007 13:45:55.648902 6373 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1007 13:45:55.648907 6373 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nF1007 13:45:55.648882 6373 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal e\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://261a778f70ae84c72b7224921a7f335d85bff09dc9ab003434f344cec3f78161\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:46:09Z\\\",\\\"message\\\":\\\"v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486086 6582 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486337 6582 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486698 6582 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486903 6582 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1007 13:46:09.487286 6582 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 13:46:09.487378 6582 handler.go:208] Removed *v1.Node event handler 2\\\\nI1007 13:46:09.488070 6582 factory.go:656] Stopping watch factory\\\\nI1007 13:46:09.505213 6582 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1007 13:46:09.505253 6582 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1007 13:46:09.505381 6582 ovnkube.go:599] Stopped ovnkube\\\\nI1007 13:46:09.505439 6582 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1007 13:46:09.505577 6582 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:46:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:10Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.123610 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ab500b1b215518cbfc4602ee0671c6f3f68464467aa392ea5bc38ffb24bd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1454006ab0b9f83bd85e57a5e7389c81aa81ea5fb4d7c86a6315839de5c37812\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xpjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:10Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.136204 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zdps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zdps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:10Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.166598 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:10Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.185806 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:10Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.202625 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:10Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.204573 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.204674 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.204689 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.204732 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.204748 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:10Z","lastTransitionTime":"2025-10-07T13:46:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.217438 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:10Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.231934 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:10Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.252542 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1cbef019f0640b5d51b398d1fc7166c1e253fd33f78f0e50f9754bb1187a6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:10Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.269986 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:10Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.285609 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:10Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.300760 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:10Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.307237 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.307276 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.307292 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.307364 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.307384 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:10Z","lastTransitionTime":"2025-10-07T13:46:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.318003 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:10Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.335005 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:10Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.411061 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.411125 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.411137 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.411158 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.411184 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:10Z","lastTransitionTime":"2025-10-07T13:46:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.514617 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.514666 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.514676 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.514693 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.514704 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:10Z","lastTransitionTime":"2025-10-07T13:46:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.590992 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.591168 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.591267 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:10 crc kubenswrapper[4959]: E1007 13:46:10.591335 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:46:42.591294869 +0000 UTC m=+84.674699224 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:46:10 crc kubenswrapper[4959]: E1007 13:46:10.591372 4959 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.591399 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:10 crc kubenswrapper[4959]: E1007 13:46:10.591455 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 13:46:42.591426382 +0000 UTC m=+84.674830807 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.591487 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:10 crc kubenswrapper[4959]: E1007 13:46:10.591622 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 13:46:10 crc kubenswrapper[4959]: E1007 13:46:10.591638 4959 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 13:46:10 crc kubenswrapper[4959]: E1007 13:46:10.591643 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 13:46:10 crc kubenswrapper[4959]: E1007 13:46:10.591661 4959 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:46:10 crc kubenswrapper[4959]: E1007 13:46:10.591708 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 13:46:42.591693749 +0000 UTC m=+84.675098114 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 13:46:10 crc kubenswrapper[4959]: E1007 13:46:10.591732 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 13:46:42.5917217 +0000 UTC m=+84.675126055 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:46:10 crc kubenswrapper[4959]: E1007 13:46:10.592240 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 13:46:10 crc kubenswrapper[4959]: E1007 13:46:10.592270 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 13:46:10 crc kubenswrapper[4959]: E1007 13:46:10.592290 4959 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:46:10 crc kubenswrapper[4959]: E1007 13:46:10.592346 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 13:46:42.592331365 +0000 UTC m=+84.675735730 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.618335 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.618407 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.618426 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.618452 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.618474 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:10Z","lastTransitionTime":"2025-10-07T13:46:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.653150 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.653243 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.653162 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:10 crc kubenswrapper[4959]: E1007 13:46:10.653372 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:10 crc kubenswrapper[4959]: E1007 13:46:10.653494 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:10 crc kubenswrapper[4959]: E1007 13:46:10.653804 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.721790 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.721844 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.721857 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.721878 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.721892 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:10Z","lastTransitionTime":"2025-10-07T13:46:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.825741 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.825811 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.825832 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.825855 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.825873 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:10Z","lastTransitionTime":"2025-10-07T13:46:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.928928 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.928972 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.928980 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.928995 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:10 crc kubenswrapper[4959]: I1007 13:46:10.929004 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:10Z","lastTransitionTime":"2025-10-07T13:46:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.012222 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5sfv5_b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b/ovnkube-controller/2.log" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.020003 4959 scope.go:117] "RemoveContainer" containerID="261a778f70ae84c72b7224921a7f335d85bff09dc9ab003434f344cec3f78161" Oct 07 13:46:11 crc kubenswrapper[4959]: E1007 13:46:11.020481 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5sfv5_openshift-ovn-kubernetes(b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.048825 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.048898 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.048924 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.048956 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.048980 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:11Z","lastTransitionTime":"2025-10-07T13:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.084887 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:11Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.111345 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:11Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.131746 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:11Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.143398 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:11Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.150638 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.150685 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.150700 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.150724 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.150741 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:11Z","lastTransitionTime":"2025-10-07T13:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.154286 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:11Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.166568 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ab500b1b215518cbfc4602ee0671c6f3f68464467aa392ea5bc38ffb24bd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1454006ab0b9f83bd85e57a5e7389c81aa81ea5fb4d7c86a6315839de5c37812\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xpjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:11Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.180050 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zdps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zdps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:11Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.195771 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:11Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.206162 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:11Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.219714 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:11Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.233366 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:11Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.247695 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1cbef019f0640b5d51b398d1fc7166c1e253fd33f78f0e50f9754bb1187a6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:11Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.253076 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.253145 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.253161 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.253180 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.253193 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:11Z","lastTransitionTime":"2025-10-07T13:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.256601 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:11Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.267636 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e325099f-e0ff-433d-bef7-6909c5c86bea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecb7da009b7ed9e5c0412125be08cf486bf4f0e74480f2843877b0816944fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6dab7bc77a409dca97cffd302d0086021bf51146e8ec69a53cbddb87a4026aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b500df391c3b1ef8f9689a04e58c19e603667507a68fb8d65ea768690425085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:11Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.279962 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:11Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.300837 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:11Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.315812 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:11Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.340600 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://261a778f70ae84c72b7224921a7f335d85bff09dc9ab003434f344cec3f78161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://261a778f70ae84c72b7224921a7f335d85bff09dc9ab003434f344cec3f78161\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:46:09Z\\\",\\\"message\\\":\\\"v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486086 6582 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486337 6582 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486698 6582 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486903 6582 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1007 13:46:09.487286 6582 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 13:46:09.487378 6582 handler.go:208] Removed *v1.Node event handler 2\\\\nI1007 13:46:09.488070 6582 factory.go:656] Stopping watch factory\\\\nI1007 13:46:09.505213 6582 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1007 13:46:09.505253 6582 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1007 13:46:09.505381 6582 ovnkube.go:599] Stopped ovnkube\\\\nI1007 13:46:09.505439 6582 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1007 13:46:09.505577 6582 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:46:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5sfv5_openshift-ovn-kubernetes(b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:11Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.355364 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.355407 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.355418 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.355435 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.355447 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:11Z","lastTransitionTime":"2025-10-07T13:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.458299 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.458329 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.458338 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.458370 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.458379 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:11Z","lastTransitionTime":"2025-10-07T13:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.561501 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.561544 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.561556 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.561572 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.561584 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:11Z","lastTransitionTime":"2025-10-07T13:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.653494 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:11 crc kubenswrapper[4959]: E1007 13:46:11.653679 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.664608 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.664670 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.664693 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.664720 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.664742 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:11Z","lastTransitionTime":"2025-10-07T13:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.767950 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.768001 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.768023 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.768052 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.768073 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:11Z","lastTransitionTime":"2025-10-07T13:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.871317 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.871392 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.871414 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.871436 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.871453 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:11Z","lastTransitionTime":"2025-10-07T13:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.973507 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.973571 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.973583 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.973599 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:11 crc kubenswrapper[4959]: I1007 13:46:11.973613 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:11Z","lastTransitionTime":"2025-10-07T13:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.075474 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.075510 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.075522 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.075537 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.075547 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:12Z","lastTransitionTime":"2025-10-07T13:46:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.177511 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.177554 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.177563 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.177578 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.177587 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:12Z","lastTransitionTime":"2025-10-07T13:46:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.279590 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.279627 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.279636 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.279651 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.279659 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:12Z","lastTransitionTime":"2025-10-07T13:46:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.382907 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.382970 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.382992 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.383019 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.383041 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:12Z","lastTransitionTime":"2025-10-07T13:46:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.486418 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.486477 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.486494 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.486521 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.486538 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:12Z","lastTransitionTime":"2025-10-07T13:46:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.589401 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.589469 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.589486 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.589511 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.589531 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:12Z","lastTransitionTime":"2025-10-07T13:46:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.652637 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.652700 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:12 crc kubenswrapper[4959]: E1007 13:46:12.652821 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.652893 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:12 crc kubenswrapper[4959]: E1007 13:46:12.652940 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:12 crc kubenswrapper[4959]: E1007 13:46:12.653157 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.692220 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.692280 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.692297 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.692322 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.692340 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:12Z","lastTransitionTime":"2025-10-07T13:46:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.795333 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.795396 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.795414 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.795470 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.795489 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:12Z","lastTransitionTime":"2025-10-07T13:46:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.898536 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.898580 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.898591 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.898606 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:12 crc kubenswrapper[4959]: I1007 13:46:12.898616 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:12Z","lastTransitionTime":"2025-10-07T13:46:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.001668 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.001697 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.001704 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.001716 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.001725 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:13Z","lastTransitionTime":"2025-10-07T13:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.104375 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.104431 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.104445 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.104461 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.104470 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:13Z","lastTransitionTime":"2025-10-07T13:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.207017 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.207084 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.207140 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.207171 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.207204 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:13Z","lastTransitionTime":"2025-10-07T13:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.309645 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.309689 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.309701 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.309718 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.309729 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:13Z","lastTransitionTime":"2025-10-07T13:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.412771 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.412825 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.412842 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.412865 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.412883 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:13Z","lastTransitionTime":"2025-10-07T13:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.516796 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.516848 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.516865 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.516888 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.516906 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:13Z","lastTransitionTime":"2025-10-07T13:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.620145 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.620204 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.620222 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.620247 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.620264 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:13Z","lastTransitionTime":"2025-10-07T13:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.652616 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:13 crc kubenswrapper[4959]: E1007 13:46:13.652799 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.724690 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.724733 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.724749 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.724771 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.724788 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:13Z","lastTransitionTime":"2025-10-07T13:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.828446 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.828495 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.828512 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.828536 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.828555 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:13Z","lastTransitionTime":"2025-10-07T13:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.931453 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.931498 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.931514 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.931537 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:13 crc kubenswrapper[4959]: I1007 13:46:13.931554 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:13Z","lastTransitionTime":"2025-10-07T13:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.033642 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.033743 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.033763 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.033793 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.033814 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:14Z","lastTransitionTime":"2025-10-07T13:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.136930 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.136997 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.137014 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.137040 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.137062 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:14Z","lastTransitionTime":"2025-10-07T13:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.240158 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.240237 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.240256 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.240282 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.240297 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:14Z","lastTransitionTime":"2025-10-07T13:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.337478 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-metrics-certs\") pod \"network-metrics-daemon-zdps5\" (UID: \"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\") " pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:14 crc kubenswrapper[4959]: E1007 13:46:14.337716 4959 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 13:46:14 crc kubenswrapper[4959]: E1007 13:46:14.337844 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-metrics-certs podName:50d4ac5c-0a16-4530-8500-9975eb2b6b5a nodeName:}" failed. No retries permitted until 2025-10-07 13:46:30.337812691 +0000 UTC m=+72.421217046 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-metrics-certs") pod "network-metrics-daemon-zdps5" (UID: "50d4ac5c-0a16-4530-8500-9975eb2b6b5a") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.344045 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.344159 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.344188 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.344221 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.344246 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:14Z","lastTransitionTime":"2025-10-07T13:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.447623 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.447708 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.447730 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.447758 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.447780 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:14Z","lastTransitionTime":"2025-10-07T13:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.552405 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.552488 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.552510 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.552539 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.552566 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:14Z","lastTransitionTime":"2025-10-07T13:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.652453 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.652485 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.652484 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:14 crc kubenswrapper[4959]: E1007 13:46:14.652584 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:14 crc kubenswrapper[4959]: E1007 13:46:14.652648 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:14 crc kubenswrapper[4959]: E1007 13:46:14.652743 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.654945 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.654974 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.654984 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.654998 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.655010 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:14Z","lastTransitionTime":"2025-10-07T13:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.757822 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.757859 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.757872 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.757893 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.757907 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:14Z","lastTransitionTime":"2025-10-07T13:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.861291 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.861339 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.861351 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.861367 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.861380 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:14Z","lastTransitionTime":"2025-10-07T13:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.964093 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.964188 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.964207 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.964231 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:14 crc kubenswrapper[4959]: I1007 13:46:14.964252 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:14Z","lastTransitionTime":"2025-10-07T13:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.067230 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.067299 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.067316 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.067341 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.067359 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:15Z","lastTransitionTime":"2025-10-07T13:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.118640 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.118698 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.118718 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.118743 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.118763 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:15Z","lastTransitionTime":"2025-10-07T13:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:15 crc kubenswrapper[4959]: E1007 13:46:15.140436 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:15Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.149547 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.150338 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.150379 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.150414 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.150439 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:15Z","lastTransitionTime":"2025-10-07T13:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:15 crc kubenswrapper[4959]: E1007 13:46:15.170208 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:15Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.175292 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.175397 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.175417 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.175440 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.175458 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:15Z","lastTransitionTime":"2025-10-07T13:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:15 crc kubenswrapper[4959]: E1007 13:46:15.195371 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:15Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.200388 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.200438 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.200455 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.200476 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.200493 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:15Z","lastTransitionTime":"2025-10-07T13:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:15 crc kubenswrapper[4959]: E1007 13:46:15.220336 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:15Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.225537 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.225692 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.225709 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.225730 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.225748 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:15Z","lastTransitionTime":"2025-10-07T13:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:15 crc kubenswrapper[4959]: E1007 13:46:15.245579 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:15Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:15 crc kubenswrapper[4959]: E1007 13:46:15.245803 4959 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.248131 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.248185 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.248207 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.248235 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.248256 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:15Z","lastTransitionTime":"2025-10-07T13:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.351192 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.351239 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.351259 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.351287 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.351312 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:15Z","lastTransitionTime":"2025-10-07T13:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.453669 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.453719 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.453735 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.453757 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.453775 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:15Z","lastTransitionTime":"2025-10-07T13:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.556882 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.556975 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.556995 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.557019 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.557041 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:15Z","lastTransitionTime":"2025-10-07T13:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.652534 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:15 crc kubenswrapper[4959]: E1007 13:46:15.652730 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.659567 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.659619 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.659642 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.659670 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.659691 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:15Z","lastTransitionTime":"2025-10-07T13:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.762075 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.762176 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.762197 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.762221 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.762240 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:15Z","lastTransitionTime":"2025-10-07T13:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.865646 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.865696 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.865714 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.865737 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.865755 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:15Z","lastTransitionTime":"2025-10-07T13:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.968318 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.969187 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.969380 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.969572 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:15 crc kubenswrapper[4959]: I1007 13:46:15.969701 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:15Z","lastTransitionTime":"2025-10-07T13:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.071997 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.072067 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.072085 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.072135 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.072152 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:16Z","lastTransitionTime":"2025-10-07T13:46:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.175373 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.175436 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.175454 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.175477 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.175494 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:16Z","lastTransitionTime":"2025-10-07T13:46:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.278726 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.278784 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.278801 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.278825 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.278845 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:16Z","lastTransitionTime":"2025-10-07T13:46:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.381915 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.381983 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.382006 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.382036 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.382057 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:16Z","lastTransitionTime":"2025-10-07T13:46:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.484649 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.484682 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.484693 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.484705 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.484714 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:16Z","lastTransitionTime":"2025-10-07T13:46:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.587913 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.587983 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.588005 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.588034 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.588055 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:16Z","lastTransitionTime":"2025-10-07T13:46:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.653450 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.653585 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:16 crc kubenswrapper[4959]: E1007 13:46:16.653608 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:16 crc kubenswrapper[4959]: E1007 13:46:16.653800 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.653851 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:16 crc kubenswrapper[4959]: E1007 13:46:16.654017 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.690857 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.690923 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.690957 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.690984 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.691005 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:16Z","lastTransitionTime":"2025-10-07T13:46:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.793334 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.793388 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.793406 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.793429 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.793446 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:16Z","lastTransitionTime":"2025-10-07T13:46:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.896418 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.896682 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.896760 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.896849 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.896932 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:16Z","lastTransitionTime":"2025-10-07T13:46:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:16 crc kubenswrapper[4959]: I1007 13:46:16.999615 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.000166 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.000247 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.000338 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.000419 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:17Z","lastTransitionTime":"2025-10-07T13:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.102779 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.102830 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.102843 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.102860 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.102873 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:17Z","lastTransitionTime":"2025-10-07T13:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.206556 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.206618 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.206632 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.206649 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.206664 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:17Z","lastTransitionTime":"2025-10-07T13:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.310343 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.310401 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.310420 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.310450 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.310472 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:17Z","lastTransitionTime":"2025-10-07T13:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.413602 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.413644 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.413657 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.413674 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.413687 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:17Z","lastTransitionTime":"2025-10-07T13:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.516790 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.516855 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.516877 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.516908 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.516929 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:17Z","lastTransitionTime":"2025-10-07T13:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.620339 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.620412 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.620432 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.620458 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.620474 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:17Z","lastTransitionTime":"2025-10-07T13:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.652748 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:17 crc kubenswrapper[4959]: E1007 13:46:17.652919 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.723201 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.723266 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.723284 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.723306 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.723323 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:17Z","lastTransitionTime":"2025-10-07T13:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.826186 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.826257 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.826280 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.826305 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.826327 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:17Z","lastTransitionTime":"2025-10-07T13:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.929437 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.929569 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.929644 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.929677 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:17 crc kubenswrapper[4959]: I1007 13:46:17.929755 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:17Z","lastTransitionTime":"2025-10-07T13:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.032874 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.032974 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.033045 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.033076 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.033176 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:18Z","lastTransitionTime":"2025-10-07T13:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.135510 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.135574 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.135590 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.135614 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.135631 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:18Z","lastTransitionTime":"2025-10-07T13:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.238168 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.238248 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.238264 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.238288 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.238345 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:18Z","lastTransitionTime":"2025-10-07T13:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.341353 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.341443 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.341467 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.341496 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.341517 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:18Z","lastTransitionTime":"2025-10-07T13:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.444550 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.444589 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.444597 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.444616 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.444626 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:18Z","lastTransitionTime":"2025-10-07T13:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.547654 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.547699 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.547710 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.547727 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.547743 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:18Z","lastTransitionTime":"2025-10-07T13:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.651227 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.651272 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.651284 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.651300 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.651313 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:18Z","lastTransitionTime":"2025-10-07T13:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.652638 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:18 crc kubenswrapper[4959]: E1007 13:46:18.652742 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.653032 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:18 crc kubenswrapper[4959]: E1007 13:46:18.653254 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.653381 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:18 crc kubenswrapper[4959]: E1007 13:46:18.653553 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.675206 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:18Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.691029 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:18Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.701482 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:18Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.711087 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:18Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.722894 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ab500b1b215518cbfc4602ee0671c6f3f68464467aa392ea5bc38ffb24bd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1454006ab0b9f83bd85e57a5e7389c81aa81ea5fb4d7c86a6315839de5c37812\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xpjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:18Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.733299 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zdps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zdps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:18Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.753576 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.753616 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.753628 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.753643 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.753656 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:18Z","lastTransitionTime":"2025-10-07T13:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.756081 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:18Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.771259 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:18Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.784427 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:18Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.798052 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:18Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.812376 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:18Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.836122 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1cbef019f0640b5d51b398d1fc7166c1e253fd33f78f0e50f9754bb1187a6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:18Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.850073 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:18Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.856700 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.856774 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.856792 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.856815 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.856833 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:18Z","lastTransitionTime":"2025-10-07T13:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.868218 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:18Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.888335 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:18Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.907440 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:18Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.940802 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://261a778f70ae84c72b7224921a7f335d85bff09dc9ab003434f344cec3f78161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://261a778f70ae84c72b7224921a7f335d85bff09dc9ab003434f344cec3f78161\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:46:09Z\\\",\\\"message\\\":\\\"v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486086 6582 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486337 6582 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486698 6582 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486903 6582 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1007 13:46:09.487286 6582 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 13:46:09.487378 6582 handler.go:208] Removed *v1.Node event handler 2\\\\nI1007 13:46:09.488070 6582 factory.go:656] Stopping watch factory\\\\nI1007 13:46:09.505213 6582 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1007 13:46:09.505253 6582 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1007 13:46:09.505381 6582 ovnkube.go:599] Stopped ovnkube\\\\nI1007 13:46:09.505439 6582 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1007 13:46:09.505577 6582 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:46:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5sfv5_openshift-ovn-kubernetes(b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:18Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.959572 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e325099f-e0ff-433d-bef7-6909c5c86bea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecb7da009b7ed9e5c0412125be08cf486bf4f0e74480f2843877b0816944fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6dab7bc77a409dca97cffd302d0086021bf51146e8ec69a53cbddb87a4026aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b500df391c3b1ef8f9689a04e58c19e603667507a68fb8d65ea768690425085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:18Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.959734 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.959767 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.959780 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.959797 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:18 crc kubenswrapper[4959]: I1007 13:46:18.959809 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:18Z","lastTransitionTime":"2025-10-07T13:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.061942 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.062005 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.062014 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.062046 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.062056 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:19Z","lastTransitionTime":"2025-10-07T13:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.165558 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.165625 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.165643 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.165668 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.165689 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:19Z","lastTransitionTime":"2025-10-07T13:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.269041 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.269184 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.269204 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.269227 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.269244 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:19Z","lastTransitionTime":"2025-10-07T13:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.372758 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.372809 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.372821 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.372839 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.372851 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:19Z","lastTransitionTime":"2025-10-07T13:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.476374 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.476459 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.476480 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.476505 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.476522 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:19Z","lastTransitionTime":"2025-10-07T13:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.579746 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.579817 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.579834 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.579862 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.579879 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:19Z","lastTransitionTime":"2025-10-07T13:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.654230 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:19 crc kubenswrapper[4959]: E1007 13:46:19.654702 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.682783 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.682822 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.682833 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.682851 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.682866 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:19Z","lastTransitionTime":"2025-10-07T13:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.784969 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.785018 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.785033 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.785055 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.785070 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:19Z","lastTransitionTime":"2025-10-07T13:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.887963 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.888017 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.888034 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.888063 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.888082 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:19Z","lastTransitionTime":"2025-10-07T13:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.991283 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.991388 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.991408 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.991429 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:19 crc kubenswrapper[4959]: I1007 13:46:19.991441 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:19Z","lastTransitionTime":"2025-10-07T13:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.094355 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.094383 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.094392 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.094403 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.094412 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:20Z","lastTransitionTime":"2025-10-07T13:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.197365 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.197436 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.197453 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.197946 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.198037 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:20Z","lastTransitionTime":"2025-10-07T13:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.301867 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.301924 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.301942 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.301969 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.301987 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:20Z","lastTransitionTime":"2025-10-07T13:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.404999 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.405052 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.405070 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.405093 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.405145 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:20Z","lastTransitionTime":"2025-10-07T13:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.507866 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.507972 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.507995 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.508025 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.508053 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:20Z","lastTransitionTime":"2025-10-07T13:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.610976 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.611028 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.611045 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.611069 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.611086 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:20Z","lastTransitionTime":"2025-10-07T13:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.653329 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.653351 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.653589 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:20 crc kubenswrapper[4959]: E1007 13:46:20.653671 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:20 crc kubenswrapper[4959]: E1007 13:46:20.653506 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:20 crc kubenswrapper[4959]: E1007 13:46:20.654335 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.715186 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.715256 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.715275 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.715300 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.715318 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:20Z","lastTransitionTime":"2025-10-07T13:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.818498 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.818562 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.818582 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.818607 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.818626 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:20Z","lastTransitionTime":"2025-10-07T13:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.923147 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.923261 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.923292 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.923327 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:20 crc kubenswrapper[4959]: I1007 13:46:20.923370 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:20Z","lastTransitionTime":"2025-10-07T13:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.026859 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.026987 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.027066 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.027151 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.027180 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:21Z","lastTransitionTime":"2025-10-07T13:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.130526 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.130567 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.130579 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.130595 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.130607 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:21Z","lastTransitionTime":"2025-10-07T13:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.234056 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.234178 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.234204 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.234234 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.234255 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:21Z","lastTransitionTime":"2025-10-07T13:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.337716 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.337769 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.337789 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.337814 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.337834 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:21Z","lastTransitionTime":"2025-10-07T13:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.441149 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.441222 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.441248 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.441277 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.441298 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:21Z","lastTransitionTime":"2025-10-07T13:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.544444 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.544517 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.544542 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.544571 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.544595 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:21Z","lastTransitionTime":"2025-10-07T13:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.646631 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.646662 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.646670 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.646683 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.646691 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:21Z","lastTransitionTime":"2025-10-07T13:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.652835 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:21 crc kubenswrapper[4959]: E1007 13:46:21.653230 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.653400 4959 scope.go:117] "RemoveContainer" containerID="261a778f70ae84c72b7224921a7f335d85bff09dc9ab003434f344cec3f78161" Oct 07 13:46:21 crc kubenswrapper[4959]: E1007 13:46:21.653587 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5sfv5_openshift-ovn-kubernetes(b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.749975 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.750078 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.750123 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.750150 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.750170 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:21Z","lastTransitionTime":"2025-10-07T13:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.854787 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.854848 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.854867 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.854891 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.854908 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:21Z","lastTransitionTime":"2025-10-07T13:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.957520 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.957577 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.957591 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.957611 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:21 crc kubenswrapper[4959]: I1007 13:46:21.957623 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:21Z","lastTransitionTime":"2025-10-07T13:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.061172 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.061225 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.061238 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.061254 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.061266 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:22Z","lastTransitionTime":"2025-10-07T13:46:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.163989 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.164057 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.164093 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.164153 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.164170 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:22Z","lastTransitionTime":"2025-10-07T13:46:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.266871 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.267337 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.267503 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.267649 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.267795 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:22Z","lastTransitionTime":"2025-10-07T13:46:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.371411 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.371779 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.371933 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.372075 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.372266 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:22Z","lastTransitionTime":"2025-10-07T13:46:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.475465 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.475704 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.475737 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.475773 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.475800 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:22Z","lastTransitionTime":"2025-10-07T13:46:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.578817 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.578878 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.578890 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.578909 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.578923 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:22Z","lastTransitionTime":"2025-10-07T13:46:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.654344 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.654402 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.654344 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:22 crc kubenswrapper[4959]: E1007 13:46:22.654469 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:22 crc kubenswrapper[4959]: E1007 13:46:22.654541 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:22 crc kubenswrapper[4959]: E1007 13:46:22.654671 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.681317 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.681356 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.681364 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.681378 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.681387 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:22Z","lastTransitionTime":"2025-10-07T13:46:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.785062 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.785151 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.785177 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.785203 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.785221 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:22Z","lastTransitionTime":"2025-10-07T13:46:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.887562 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.887607 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.887617 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.887640 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.887670 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:22Z","lastTransitionTime":"2025-10-07T13:46:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.990333 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.990396 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.990413 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.990438 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:22 crc kubenswrapper[4959]: I1007 13:46:22.990455 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:22Z","lastTransitionTime":"2025-10-07T13:46:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.092594 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.092664 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.092680 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.092703 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.092720 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:23Z","lastTransitionTime":"2025-10-07T13:46:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.196285 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.196333 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.196347 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.196365 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.196377 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:23Z","lastTransitionTime":"2025-10-07T13:46:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.299133 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.299209 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.299226 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.299246 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.299287 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:23Z","lastTransitionTime":"2025-10-07T13:46:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.401188 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.401261 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.401280 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.401306 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.401324 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:23Z","lastTransitionTime":"2025-10-07T13:46:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.503665 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.503707 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.503717 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.503733 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.503759 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:23Z","lastTransitionTime":"2025-10-07T13:46:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.606523 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.606570 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.606582 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.606598 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.606610 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:23Z","lastTransitionTime":"2025-10-07T13:46:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.652400 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:23 crc kubenswrapper[4959]: E1007 13:46:23.652516 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.709064 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.709232 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.709252 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.709274 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.709318 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:23Z","lastTransitionTime":"2025-10-07T13:46:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.811987 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.812035 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.812048 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.812064 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.812077 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:23Z","lastTransitionTime":"2025-10-07T13:46:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.914450 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.914500 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.914508 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.914525 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:23 crc kubenswrapper[4959]: I1007 13:46:23.914539 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:23Z","lastTransitionTime":"2025-10-07T13:46:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.017268 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.017309 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.017318 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.017332 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.017357 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:24Z","lastTransitionTime":"2025-10-07T13:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.119367 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.119444 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.119470 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.119501 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.119523 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:24Z","lastTransitionTime":"2025-10-07T13:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.222415 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.222476 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.222492 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.222514 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.222531 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:24Z","lastTransitionTime":"2025-10-07T13:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.325130 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.325183 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.325194 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.325211 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.325227 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:24Z","lastTransitionTime":"2025-10-07T13:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.427656 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.427694 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.427705 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.427720 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.427730 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:24Z","lastTransitionTime":"2025-10-07T13:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.529838 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.529897 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.529918 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.529943 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.529960 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:24Z","lastTransitionTime":"2025-10-07T13:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.634541 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.634614 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.634637 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.634661 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.634678 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:24Z","lastTransitionTime":"2025-10-07T13:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.653379 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:24 crc kubenswrapper[4959]: E1007 13:46:24.653559 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.653970 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:24 crc kubenswrapper[4959]: E1007 13:46:24.654175 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.654591 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:24 crc kubenswrapper[4959]: E1007 13:46:24.654735 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.737066 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.737138 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.737155 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.737176 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.737193 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:24Z","lastTransitionTime":"2025-10-07T13:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.840599 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.840657 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.840672 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.840695 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.840710 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:24Z","lastTransitionTime":"2025-10-07T13:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.942444 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.942514 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.942534 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.942557 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:24 crc kubenswrapper[4959]: I1007 13:46:24.942575 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:24Z","lastTransitionTime":"2025-10-07T13:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.045376 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.045493 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.045520 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.045567 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.045594 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:25Z","lastTransitionTime":"2025-10-07T13:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.148331 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.148385 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.148400 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.148418 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.148429 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:25Z","lastTransitionTime":"2025-10-07T13:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.252071 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.252159 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.252178 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.252202 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.252219 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:25Z","lastTransitionTime":"2025-10-07T13:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.284811 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.284856 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.284868 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.284886 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.284900 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:25Z","lastTransitionTime":"2025-10-07T13:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:25 crc kubenswrapper[4959]: E1007 13:46:25.304657 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:25Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.309630 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.309710 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.309735 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.309760 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.309780 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:25Z","lastTransitionTime":"2025-10-07T13:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:25 crc kubenswrapper[4959]: E1007 13:46:25.331679 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:25Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.335417 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.335455 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.335467 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.335483 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.335494 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:25Z","lastTransitionTime":"2025-10-07T13:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:25 crc kubenswrapper[4959]: E1007 13:46:25.352596 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:25Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.356615 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.356645 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.356655 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.356669 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.356687 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:25Z","lastTransitionTime":"2025-10-07T13:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:25 crc kubenswrapper[4959]: E1007 13:46:25.373202 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:25Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.377185 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.377234 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.377247 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.377264 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.377275 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:25Z","lastTransitionTime":"2025-10-07T13:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:25 crc kubenswrapper[4959]: E1007 13:46:25.395922 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:25Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:25 crc kubenswrapper[4959]: E1007 13:46:25.396180 4959 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.398046 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.398095 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.398163 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.398190 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.398207 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:25Z","lastTransitionTime":"2025-10-07T13:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.500994 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.501045 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.501062 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.501085 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.501126 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:25Z","lastTransitionTime":"2025-10-07T13:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.603292 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.603493 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.603594 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.603673 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.603743 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:25Z","lastTransitionTime":"2025-10-07T13:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.653370 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:25 crc kubenswrapper[4959]: E1007 13:46:25.653558 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.706664 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.706770 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.706781 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.706795 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.706804 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:25Z","lastTransitionTime":"2025-10-07T13:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.816246 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.816303 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.816321 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.816343 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.816360 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:25Z","lastTransitionTime":"2025-10-07T13:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.918974 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.919236 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.919295 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.919354 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:25 crc kubenswrapper[4959]: I1007 13:46:25.919410 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:25Z","lastTransitionTime":"2025-10-07T13:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.021708 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.021925 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.022034 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.022068 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.022092 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:26Z","lastTransitionTime":"2025-10-07T13:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.124909 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.125153 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.125269 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.125345 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.125414 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:26Z","lastTransitionTime":"2025-10-07T13:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.228481 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.228515 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.228534 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.228549 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.228559 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:26Z","lastTransitionTime":"2025-10-07T13:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.330879 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.330934 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.330950 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.330974 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.330991 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:26Z","lastTransitionTime":"2025-10-07T13:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.439931 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.439981 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.440000 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.440024 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.440041 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:26Z","lastTransitionTime":"2025-10-07T13:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.543421 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.543484 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.543506 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.543536 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.543559 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:26Z","lastTransitionTime":"2025-10-07T13:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.646811 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.646852 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.646864 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.646883 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.646898 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:26Z","lastTransitionTime":"2025-10-07T13:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.653368 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.653397 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.653368 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:26 crc kubenswrapper[4959]: E1007 13:46:26.653482 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:26 crc kubenswrapper[4959]: E1007 13:46:26.653543 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:26 crc kubenswrapper[4959]: E1007 13:46:26.653678 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.748601 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.748632 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.748640 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.748655 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.748664 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:26Z","lastTransitionTime":"2025-10-07T13:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.851165 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.851232 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.851255 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.851278 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.851295 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:26Z","lastTransitionTime":"2025-10-07T13:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.954421 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.954483 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.954506 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.954535 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:26 crc kubenswrapper[4959]: I1007 13:46:26.954559 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:26Z","lastTransitionTime":"2025-10-07T13:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.057977 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.058018 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.058029 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.058045 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.058055 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:27Z","lastTransitionTime":"2025-10-07T13:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.160659 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.160701 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.160719 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.160740 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.160758 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:27Z","lastTransitionTime":"2025-10-07T13:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.263974 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.264035 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.264059 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.264088 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.264152 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:27Z","lastTransitionTime":"2025-10-07T13:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.366850 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.366882 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.366893 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.366907 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.366918 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:27Z","lastTransitionTime":"2025-10-07T13:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.470044 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.470093 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.470121 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.470137 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.470148 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:27Z","lastTransitionTime":"2025-10-07T13:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.572861 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.572939 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.572963 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.572993 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.573015 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:27Z","lastTransitionTime":"2025-10-07T13:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.652909 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:27 crc kubenswrapper[4959]: E1007 13:46:27.653792 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.675554 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.675693 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.675712 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.675737 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.675755 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:27Z","lastTransitionTime":"2025-10-07T13:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.778558 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.778655 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.778668 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.778687 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.778700 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:27Z","lastTransitionTime":"2025-10-07T13:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.881415 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.881484 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.881504 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.881531 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.881550 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:27Z","lastTransitionTime":"2025-10-07T13:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.984159 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.984282 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.984381 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.984407 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:27 crc kubenswrapper[4959]: I1007 13:46:27.984424 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:27Z","lastTransitionTime":"2025-10-07T13:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.086156 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.086201 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.086209 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.086226 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.086238 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:28Z","lastTransitionTime":"2025-10-07T13:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.188724 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.188766 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.188779 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.188796 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.188811 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:28Z","lastTransitionTime":"2025-10-07T13:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.290514 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.290563 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.290579 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.290602 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.290620 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:28Z","lastTransitionTime":"2025-10-07T13:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.392400 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.392458 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.392477 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.392500 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.392517 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:28Z","lastTransitionTime":"2025-10-07T13:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.494727 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.495346 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.495539 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.495723 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.495919 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:28Z","lastTransitionTime":"2025-10-07T13:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.598502 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.598795 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.598971 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.599144 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.599273 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:28Z","lastTransitionTime":"2025-10-07T13:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.652618 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:28 crc kubenswrapper[4959]: E1007 13:46:28.652798 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.652844 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.652937 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:28 crc kubenswrapper[4959]: E1007 13:46:28.653058 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:28 crc kubenswrapper[4959]: E1007 13:46:28.653194 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.667695 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:28Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.680409 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:28Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.692435 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1cbef019f0640b5d51b398d1fc7166c1e253fd33f78f0e50f9754bb1187a6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:28Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.701384 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.701655 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.701790 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.701921 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.701973 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:28Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.702140 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:28Z","lastTransitionTime":"2025-10-07T13:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.719662 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:28Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.730972 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:28Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.740294 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:28Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.758705 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://261a778f70ae84c72b7224921a7f335d85bff09dc9ab003434f344cec3f78161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://261a778f70ae84c72b7224921a7f335d85bff09dc9ab003434f344cec3f78161\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:46:09Z\\\",\\\"message\\\":\\\"v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486086 6582 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486337 6582 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486698 6582 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486903 6582 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1007 13:46:09.487286 6582 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 13:46:09.487378 6582 handler.go:208] Removed *v1.Node event handler 2\\\\nI1007 13:46:09.488070 6582 factory.go:656] Stopping watch factory\\\\nI1007 13:46:09.505213 6582 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1007 13:46:09.505253 6582 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1007 13:46:09.505381 6582 ovnkube.go:599] Stopped ovnkube\\\\nI1007 13:46:09.505439 6582 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1007 13:46:09.505577 6582 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:46:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5sfv5_openshift-ovn-kubernetes(b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:28Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.770872 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e325099f-e0ff-433d-bef7-6909c5c86bea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecb7da009b7ed9e5c0412125be08cf486bf4f0e74480f2843877b0816944fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6dab7bc77a409dca97cffd302d0086021bf51146e8ec69a53cbddb87a4026aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b500df391c3b1ef8f9689a04e58c19e603667507a68fb8d65ea768690425085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:28Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.787542 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:28Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.800858 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:28Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.803970 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.804007 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.804016 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.804029 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.804038 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:28Z","lastTransitionTime":"2025-10-07T13:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.809790 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:28Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.821879 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:28Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.831490 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ab500b1b215518cbfc4602ee0671c6f3f68464467aa392ea5bc38ffb24bd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1454006ab0b9f83bd85e57a5e7389c81aa81ea5fb4d7c86a6315839de5c37812\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xpjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:28Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.843818 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zdps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zdps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:28Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.875928 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:28Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.887227 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:28Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.901978 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:28Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.906475 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.906522 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.906536 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.906554 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:28 crc kubenswrapper[4959]: I1007 13:46:28.906598 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:28Z","lastTransitionTime":"2025-10-07T13:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.009790 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.009853 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.009869 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.009894 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.009911 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:29Z","lastTransitionTime":"2025-10-07T13:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.112846 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.112905 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.112922 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.112945 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.112958 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:29Z","lastTransitionTime":"2025-10-07T13:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.217468 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.217510 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.217519 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.217534 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.217545 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:29Z","lastTransitionTime":"2025-10-07T13:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.320330 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.320359 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.320367 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.320380 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.320389 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:29Z","lastTransitionTime":"2025-10-07T13:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.422485 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.422588 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.422617 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.422648 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.422672 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:29Z","lastTransitionTime":"2025-10-07T13:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.524802 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.524853 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.524865 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.524884 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.524897 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:29Z","lastTransitionTime":"2025-10-07T13:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.627609 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.627639 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.627647 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.627661 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.627670 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:29Z","lastTransitionTime":"2025-10-07T13:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.652953 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:29 crc kubenswrapper[4959]: E1007 13:46:29.653128 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.730830 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.730926 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.730946 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.730970 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.730988 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:29Z","lastTransitionTime":"2025-10-07T13:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.833437 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.833501 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.833519 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.833545 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.833564 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:29Z","lastTransitionTime":"2025-10-07T13:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.936576 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.936615 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.936641 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.936661 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:29 crc kubenswrapper[4959]: I1007 13:46:29.936673 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:29Z","lastTransitionTime":"2025-10-07T13:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.039477 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.039511 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.039550 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.039570 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.039582 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:30Z","lastTransitionTime":"2025-10-07T13:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.141467 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.141501 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.141510 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.141524 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.141534 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:30Z","lastTransitionTime":"2025-10-07T13:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.243593 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.243625 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.243634 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.243650 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.243658 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:30Z","lastTransitionTime":"2025-10-07T13:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.345813 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.345850 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.345859 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.345874 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.345884 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:30Z","lastTransitionTime":"2025-10-07T13:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.405301 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-metrics-certs\") pod \"network-metrics-daemon-zdps5\" (UID: \"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\") " pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:30 crc kubenswrapper[4959]: E1007 13:46:30.405539 4959 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 13:46:30 crc kubenswrapper[4959]: E1007 13:46:30.405661 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-metrics-certs podName:50d4ac5c-0a16-4530-8500-9975eb2b6b5a nodeName:}" failed. No retries permitted until 2025-10-07 13:47:02.405633494 +0000 UTC m=+104.489037859 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-metrics-certs") pod "network-metrics-daemon-zdps5" (UID: "50d4ac5c-0a16-4530-8500-9975eb2b6b5a") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.448341 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.448382 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.448391 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.448404 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.448413 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:30Z","lastTransitionTime":"2025-10-07T13:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.551191 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.551232 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.551241 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.551256 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.551266 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:30Z","lastTransitionTime":"2025-10-07T13:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.652540 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.652602 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.652718 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:30 crc kubenswrapper[4959]: E1007 13:46:30.652707 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:30 crc kubenswrapper[4959]: E1007 13:46:30.652798 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:30 crc kubenswrapper[4959]: E1007 13:46:30.652976 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.653379 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.653431 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.653455 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.653472 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.653483 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:30Z","lastTransitionTime":"2025-10-07T13:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.755823 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.755876 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.755888 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.755906 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.755920 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:30Z","lastTransitionTime":"2025-10-07T13:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.858534 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.858567 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.858576 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.858589 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.858597 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:30Z","lastTransitionTime":"2025-10-07T13:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.960875 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.961432 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.961528 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.961897 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:30 crc kubenswrapper[4959]: I1007 13:46:30.961948 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:30Z","lastTransitionTime":"2025-10-07T13:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.064261 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.064300 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.064308 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.064321 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.064330 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:31Z","lastTransitionTime":"2025-10-07T13:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.166199 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.166236 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.166247 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.166264 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.166276 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:31Z","lastTransitionTime":"2025-10-07T13:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.268294 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.268348 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.268377 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.268395 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.268408 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:31Z","lastTransitionTime":"2025-10-07T13:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.371117 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.371146 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.371154 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.371166 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.371175 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:31Z","lastTransitionTime":"2025-10-07T13:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.472771 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.472828 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.472838 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.472852 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.472885 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:31Z","lastTransitionTime":"2025-10-07T13:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.574430 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.574485 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.574497 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.574512 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.574522 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:31Z","lastTransitionTime":"2025-10-07T13:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.653195 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:31 crc kubenswrapper[4959]: E1007 13:46:31.653337 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.668650 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.676590 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.676626 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.676642 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.676658 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.676670 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:31Z","lastTransitionTime":"2025-10-07T13:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.779364 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.779430 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.779456 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.779487 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.779511 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:31Z","lastTransitionTime":"2025-10-07T13:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.882610 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.882749 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.882771 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.882806 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.882832 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:31Z","lastTransitionTime":"2025-10-07T13:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.986151 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.986206 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.986226 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.986250 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:31 crc kubenswrapper[4959]: I1007 13:46:31.986269 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:31Z","lastTransitionTime":"2025-10-07T13:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.088797 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.088859 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.088872 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.088895 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.088906 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:32Z","lastTransitionTime":"2025-10-07T13:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.090312 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4mzl6_2003ec8f-74a0-47bc-8998-0326bfff6e7f/kube-multus/0.log" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.090404 4959 generic.go:334] "Generic (PLEG): container finished" podID="2003ec8f-74a0-47bc-8998-0326bfff6e7f" containerID="9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb" exitCode=1 Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.091175 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4mzl6" event={"ID":"2003ec8f-74a0-47bc-8998-0326bfff6e7f","Type":"ContainerDied","Data":"9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb"} Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.091591 4959 scope.go:117] "RemoveContainer" containerID="9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.114077 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e325099f-e0ff-433d-bef7-6909c5c86bea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecb7da009b7ed9e5c0412125be08cf486bf4f0e74480f2843877b0816944fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6dab7bc77a409dca97cffd302d0086021bf51146e8ec69a53cbddb87a4026aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b500df391c3b1ef8f9689a04e58c19e603667507a68fb8d65ea768690425085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:32Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.137503 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:32Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.158366 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:32Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.179291 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:32Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.192314 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.192363 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.192375 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.192395 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.192407 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:32Z","lastTransitionTime":"2025-10-07T13:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.210287 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://261a778f70ae84c72b7224921a7f335d85bff09dc9ab003434f344cec3f78161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://261a778f70ae84c72b7224921a7f335d85bff09dc9ab003434f344cec3f78161\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:46:09Z\\\",\\\"message\\\":\\\"v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486086 6582 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486337 6582 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486698 6582 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486903 6582 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1007 13:46:09.487286 6582 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 13:46:09.487378 6582 handler.go:208] Removed *v1.Node event handler 2\\\\nI1007 13:46:09.488070 6582 factory.go:656] Stopping watch factory\\\\nI1007 13:46:09.505213 6582 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1007 13:46:09.505253 6582 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1007 13:46:09.505381 6582 ovnkube.go:599] Stopped ovnkube\\\\nI1007 13:46:09.505439 6582 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1007 13:46:09.505577 6582 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:46:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5sfv5_openshift-ovn-kubernetes(b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:32Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.226052 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ab500b1b215518cbfc4602ee0671c6f3f68464467aa392ea5bc38ffb24bd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1454006ab0b9f83bd85e57a5e7389c81aa81ea5fb4d7c86a6315839de5c37812\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xpjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:32Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.241654 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zdps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zdps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:32Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.269602 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:32Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.288071 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:32Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.294718 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.294762 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.294780 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.294802 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.294816 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:32Z","lastTransitionTime":"2025-10-07T13:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.301799 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:32Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.314704 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:32Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.330453 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:32Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.351785 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1cbef019f0640b5d51b398d1fc7166c1e253fd33f78f0e50f9754bb1187a6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:32Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.366739 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:32Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.381327 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1ac79b-8fe3-40cd-925d-d840f0995abf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae966401c1e804da8649ca9c014cfc2eed8dbb12a0576cf90b6c7a3526513601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6d82be64b945bd47d6149c695f43152998644d31f0103bfff861199d5efb191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6d82be64b945bd47d6149c695f43152998644d31f0103bfff861199d5efb191\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:32Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.400154 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.400251 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.400274 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.400303 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.400332 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:32Z","lastTransitionTime":"2025-10-07T13:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.406430 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:32Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.424639 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:32Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.440254 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:32Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.456197 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:46:31Z\\\",\\\"message\\\":\\\"2025-10-07T13:45:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_19b999ee-ab24-4a80-bd1e-5817a041f450\\\\n2025-10-07T13:45:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_19b999ee-ab24-4a80-bd1e-5817a041f450 to /host/opt/cni/bin/\\\\n2025-10-07T13:45:46Z [verbose] multus-daemon started\\\\n2025-10-07T13:45:46Z [verbose] Readiness Indicator file check\\\\n2025-10-07T13:46:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:32Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.503369 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.503408 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.503440 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.503457 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.503467 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:32Z","lastTransitionTime":"2025-10-07T13:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.606341 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.606391 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.606402 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.606422 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.606435 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:32Z","lastTransitionTime":"2025-10-07T13:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.653397 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.653481 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:32 crc kubenswrapper[4959]: E1007 13:46:32.653533 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.653413 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:32 crc kubenswrapper[4959]: E1007 13:46:32.653605 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:32 crc kubenswrapper[4959]: E1007 13:46:32.653648 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.709056 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.709115 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.709128 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.709144 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.709155 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:32Z","lastTransitionTime":"2025-10-07T13:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.812070 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.812154 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.812172 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.812196 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.812213 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:32Z","lastTransitionTime":"2025-10-07T13:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.916146 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.916206 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.916226 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.916249 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:32 crc kubenswrapper[4959]: I1007 13:46:32.916267 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:32Z","lastTransitionTime":"2025-10-07T13:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.020459 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.020573 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.020593 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.020623 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.020645 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:33Z","lastTransitionTime":"2025-10-07T13:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.097440 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4mzl6_2003ec8f-74a0-47bc-8998-0326bfff6e7f/kube-multus/0.log" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.097546 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4mzl6" event={"ID":"2003ec8f-74a0-47bc-8998-0326bfff6e7f","Type":"ContainerStarted","Data":"6898f5d8eb8c5a06853350d8189d1c393cef00392a0c6240b449dbb8617d9f2f"} Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.113334 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e325099f-e0ff-433d-bef7-6909c5c86bea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecb7da009b7ed9e5c0412125be08cf486bf4f0e74480f2843877b0816944fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6dab7bc77a409dca97cffd302d0086021bf51146e8ec69a53cbddb87a4026aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b500df391c3b1ef8f9689a04e58c19e603667507a68fb8d65ea768690425085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:33Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.128938 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.129014 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.129039 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.129075 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.129137 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:33Z","lastTransitionTime":"2025-10-07T13:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.131948 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:33Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.150786 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:33Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.168481 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:33Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.201006 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://261a778f70ae84c72b7224921a7f335d85bff09dc9ab003434f344cec3f78161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://261a778f70ae84c72b7224921a7f335d85bff09dc9ab003434f344cec3f78161\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:46:09Z\\\",\\\"message\\\":\\\"v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486086 6582 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486337 6582 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486698 6582 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486903 6582 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1007 13:46:09.487286 6582 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 13:46:09.487378 6582 handler.go:208] Removed *v1.Node event handler 2\\\\nI1007 13:46:09.488070 6582 factory.go:656] Stopping watch factory\\\\nI1007 13:46:09.505213 6582 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1007 13:46:09.505253 6582 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1007 13:46:09.505381 6582 ovnkube.go:599] Stopped ovnkube\\\\nI1007 13:46:09.505439 6582 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1007 13:46:09.505577 6582 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:46:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5sfv5_openshift-ovn-kubernetes(b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:33Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.218161 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zdps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zdps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:33Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.234928 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.234970 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.234987 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.235010 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.235027 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:33Z","lastTransitionTime":"2025-10-07T13:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.253688 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:33Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.271899 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:33Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.290343 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:33Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.302475 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:33Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.315651 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:33Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.331497 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ab500b1b215518cbfc4602ee0671c6f3f68464467aa392ea5bc38ffb24bd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1454006ab0b9f83bd85e57a5e7389c81aa81ea5fb4d7c86a6315839de5c37812\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xpjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:33Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.337451 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.337488 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.337504 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.337527 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.337544 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:33Z","lastTransitionTime":"2025-10-07T13:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.342840 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:33Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.352669 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1ac79b-8fe3-40cd-925d-d840f0995abf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae966401c1e804da8649ca9c014cfc2eed8dbb12a0576cf90b6c7a3526513601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6d82be64b945bd47d6149c695f43152998644d31f0103bfff861199d5efb191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6d82be64b945bd47d6149c695f43152998644d31f0103bfff861199d5efb191\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:33Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.385624 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:33Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.398436 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:33Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.414242 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:33Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.432497 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6898f5d8eb8c5a06853350d8189d1c393cef00392a0c6240b449dbb8617d9f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:46:31Z\\\",\\\"message\\\":\\\"2025-10-07T13:45:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_19b999ee-ab24-4a80-bd1e-5817a041f450\\\\n2025-10-07T13:45:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_19b999ee-ab24-4a80-bd1e-5817a041f450 to /host/opt/cni/bin/\\\\n2025-10-07T13:45:46Z [verbose] multus-daemon started\\\\n2025-10-07T13:45:46Z [verbose] Readiness Indicator file check\\\\n2025-10-07T13:46:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:46:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:33Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.440720 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.440796 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.440820 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.440854 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.440877 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:33Z","lastTransitionTime":"2025-10-07T13:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.453870 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1cbef019f0640b5d51b398d1fc7166c1e253fd33f78f0e50f9754bb1187a6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:33Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.543624 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.543678 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.543695 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.543764 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.543783 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:33Z","lastTransitionTime":"2025-10-07T13:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.645960 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.646008 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.646021 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.646038 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.646052 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:33Z","lastTransitionTime":"2025-10-07T13:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.652445 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:33 crc kubenswrapper[4959]: E1007 13:46:33.652622 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.749701 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.749756 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.749773 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.749797 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.749814 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:33Z","lastTransitionTime":"2025-10-07T13:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.852502 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.852578 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.852596 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.852621 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.852638 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:33Z","lastTransitionTime":"2025-10-07T13:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.954576 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.954649 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.954669 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.954696 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:33 crc kubenswrapper[4959]: I1007 13:46:33.954715 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:33Z","lastTransitionTime":"2025-10-07T13:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.057116 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.057166 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.057174 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.057220 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.057231 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:34Z","lastTransitionTime":"2025-10-07T13:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.159889 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.159944 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.159963 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.159986 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.160006 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:34Z","lastTransitionTime":"2025-10-07T13:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.262493 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.262521 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.262529 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.262542 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.262550 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:34Z","lastTransitionTime":"2025-10-07T13:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.365608 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.365662 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.365682 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.365709 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.365730 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:34Z","lastTransitionTime":"2025-10-07T13:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.468501 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.468848 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.469244 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.469468 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.469647 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:34Z","lastTransitionTime":"2025-10-07T13:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.572264 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.573131 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.573296 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.573446 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.573570 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:34Z","lastTransitionTime":"2025-10-07T13:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.658416 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:34 crc kubenswrapper[4959]: E1007 13:46:34.658645 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.658433 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:34 crc kubenswrapper[4959]: E1007 13:46:34.659087 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.658930 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:34 crc kubenswrapper[4959]: E1007 13:46:34.659746 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.676565 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.676643 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.676666 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.676700 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.676724 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:34Z","lastTransitionTime":"2025-10-07T13:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.778792 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.779207 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.779381 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.779542 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.779681 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:34Z","lastTransitionTime":"2025-10-07T13:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.882598 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.882657 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.882676 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.882706 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.882724 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:34Z","lastTransitionTime":"2025-10-07T13:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.985446 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.985520 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.985545 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.985574 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:34 crc kubenswrapper[4959]: I1007 13:46:34.985599 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:34Z","lastTransitionTime":"2025-10-07T13:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.089155 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.089212 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.089228 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.089255 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.089273 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:35Z","lastTransitionTime":"2025-10-07T13:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.192251 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.192309 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.192327 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.192350 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.192368 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:35Z","lastTransitionTime":"2025-10-07T13:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.295391 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.295442 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.295461 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.295484 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.295502 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:35Z","lastTransitionTime":"2025-10-07T13:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.399140 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.399206 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.399224 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.399248 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.399264 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:35Z","lastTransitionTime":"2025-10-07T13:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.471965 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.472047 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.472076 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.472133 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.472152 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:35Z","lastTransitionTime":"2025-10-07T13:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:35 crc kubenswrapper[4959]: E1007 13:46:35.493412 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:35Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.498820 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.498875 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.498896 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.498923 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.498940 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:35Z","lastTransitionTime":"2025-10-07T13:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:35 crc kubenswrapper[4959]: E1007 13:46:35.518636 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:35Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.523549 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.523618 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.523638 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.523665 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.523688 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:35Z","lastTransitionTime":"2025-10-07T13:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:35 crc kubenswrapper[4959]: E1007 13:46:35.546933 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:35Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.553659 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.553718 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.553734 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.553758 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.553781 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:35Z","lastTransitionTime":"2025-10-07T13:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:35 crc kubenswrapper[4959]: E1007 13:46:35.581715 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:35Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.587083 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.587158 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.587176 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.587201 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.587219 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:35Z","lastTransitionTime":"2025-10-07T13:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:35 crc kubenswrapper[4959]: E1007 13:46:35.607401 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:35Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:35 crc kubenswrapper[4959]: E1007 13:46:35.607616 4959 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.609680 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.609743 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.609763 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.609789 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.609809 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:35Z","lastTransitionTime":"2025-10-07T13:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.653403 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:35 crc kubenswrapper[4959]: E1007 13:46:35.653617 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.654672 4959 scope.go:117] "RemoveContainer" containerID="261a778f70ae84c72b7224921a7f335d85bff09dc9ab003434f344cec3f78161" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.713284 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.713392 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.713413 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.713487 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.713542 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:35Z","lastTransitionTime":"2025-10-07T13:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.816575 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.816626 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.816643 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.816666 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.816685 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:35Z","lastTransitionTime":"2025-10-07T13:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.919719 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.919761 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.919777 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.919802 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:35 crc kubenswrapper[4959]: I1007 13:46:35.919818 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:35Z","lastTransitionTime":"2025-10-07T13:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.022083 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.022148 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.022162 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.022181 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.022196 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:36Z","lastTransitionTime":"2025-10-07T13:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.110792 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5sfv5_b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b/ovnkube-controller/2.log" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.113682 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" event={"ID":"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b","Type":"ContainerStarted","Data":"fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69"} Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.114935 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.124379 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.124419 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.124430 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.124448 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.124460 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:36Z","lastTransitionTime":"2025-10-07T13:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.138760 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1ac79b-8fe3-40cd-925d-d840f0995abf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae966401c1e804da8649ca9c014cfc2eed8dbb12a0576cf90b6c7a3526513601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6d82be64b945bd47d6149c695f43152998644d31f0103bfff861199d5efb191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6d82be64b945bd47d6149c695f43152998644d31f0103bfff861199d5efb191\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:36Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.153542 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:36Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.166154 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:36Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.177832 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:36Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.189791 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6898f5d8eb8c5a06853350d8189d1c393cef00392a0c6240b449dbb8617d9f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:46:31Z\\\",\\\"message\\\":\\\"2025-10-07T13:45:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_19b999ee-ab24-4a80-bd1e-5817a041f450\\\\n2025-10-07T13:45:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_19b999ee-ab24-4a80-bd1e-5817a041f450 to /host/opt/cni/bin/\\\\n2025-10-07T13:45:46Z [verbose] multus-daemon started\\\\n2025-10-07T13:45:46Z [verbose] Readiness Indicator file check\\\\n2025-10-07T13:46:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:46:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:36Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.206942 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1cbef019f0640b5d51b398d1fc7166c1e253fd33f78f0e50f9754bb1187a6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:36Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.218026 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:36Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.226889 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.226927 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.226938 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.226955 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.226967 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:36Z","lastTransitionTime":"2025-10-07T13:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.229590 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e325099f-e0ff-433d-bef7-6909c5c86bea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecb7da009b7ed9e5c0412125be08cf486bf4f0e74480f2843877b0816944fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6dab7bc77a409dca97cffd302d0086021bf51146e8ec69a53cbddb87a4026aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b500df391c3b1ef8f9689a04e58c19e603667507a68fb8d65ea768690425085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:36Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.243887 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:36Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.269233 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:36Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.285148 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:36Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.306236 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://261a778f70ae84c72b7224921a7f335d85bff09dc9ab003434f344cec3f78161\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:46:09Z\\\",\\\"message\\\":\\\"v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486086 6582 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486337 6582 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486698 6582 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486903 6582 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1007 13:46:09.487286 6582 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 13:46:09.487378 6582 handler.go:208] Removed *v1.Node event handler 2\\\\nI1007 13:46:09.488070 6582 factory.go:656] Stopping watch factory\\\\nI1007 13:46:09.505213 6582 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1007 13:46:09.505253 6582 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1007 13:46:09.505381 6582 ovnkube.go:599] Stopped ovnkube\\\\nI1007 13:46:09.505439 6582 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1007 13:46:09.505577 6582 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:46:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:36Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.328247 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:36Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.329405 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.329449 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.329463 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.329482 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.329495 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:36Z","lastTransitionTime":"2025-10-07T13:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.344366 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:36Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.357483 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:36Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.373143 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:36Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.386340 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:36Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.401803 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ab500b1b215518cbfc4602ee0671c6f3f68464467aa392ea5bc38ffb24bd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1454006ab0b9f83bd85e57a5e7389c81aa81ea5fb4d7c86a6315839de5c37812\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xpjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:36Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.414664 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zdps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zdps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:36Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.431191 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.431234 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.431244 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.431261 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.431273 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:36Z","lastTransitionTime":"2025-10-07T13:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.536873 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.536944 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.536962 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.536988 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.537006 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:36Z","lastTransitionTime":"2025-10-07T13:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.639708 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.639743 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.639754 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.639770 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.639782 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:36Z","lastTransitionTime":"2025-10-07T13:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.653316 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.653447 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:36 crc kubenswrapper[4959]: E1007 13:46:36.653589 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.653630 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:36 crc kubenswrapper[4959]: E1007 13:46:36.653733 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:36 crc kubenswrapper[4959]: E1007 13:46:36.653835 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.743043 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.743112 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.743127 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.743165 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.743181 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:36Z","lastTransitionTime":"2025-10-07T13:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.846086 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.846166 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.846184 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.846257 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.846274 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:36Z","lastTransitionTime":"2025-10-07T13:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.949208 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.949247 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.949262 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.949281 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:36 crc kubenswrapper[4959]: I1007 13:46:36.949297 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:36Z","lastTransitionTime":"2025-10-07T13:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.052059 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.052143 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.052164 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.052193 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.052213 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:37Z","lastTransitionTime":"2025-10-07T13:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.121077 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5sfv5_b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b/ovnkube-controller/3.log" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.122190 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5sfv5_b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b/ovnkube-controller/2.log" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.127156 4959 generic.go:334] "Generic (PLEG): container finished" podID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerID="fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69" exitCode=1 Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.127206 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" event={"ID":"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b","Type":"ContainerDied","Data":"fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69"} Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.127305 4959 scope.go:117] "RemoveContainer" containerID="261a778f70ae84c72b7224921a7f335d85bff09dc9ab003434f344cec3f78161" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.128156 4959 scope.go:117] "RemoveContainer" containerID="fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69" Oct 07 13:46:37 crc kubenswrapper[4959]: E1007 13:46:37.128412 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5sfv5_openshift-ovn-kubernetes(b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.148586 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:37Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.154936 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.154986 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.155008 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.155041 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.155062 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:37Z","lastTransitionTime":"2025-10-07T13:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.170989 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:37Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.192528 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:37Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.223405 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://261a778f70ae84c72b7224921a7f335d85bff09dc9ab003434f344cec3f78161\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:46:09Z\\\",\\\"message\\\":\\\"v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486086 6582 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486337 6582 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486698 6582 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1007 13:46:09.486903 6582 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1007 13:46:09.487286 6582 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1007 13:46:09.487378 6582 handler.go:208] Removed *v1.Node event handler 2\\\\nI1007 13:46:09.488070 6582 factory.go:656] Stopping watch factory\\\\nI1007 13:46:09.505213 6582 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1007 13:46:09.505253 6582 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1007 13:46:09.505381 6582 ovnkube.go:599] Stopped ovnkube\\\\nI1007 13:46:09.505439 6582 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1007 13:46:09.505577 6582 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:46:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:46:36Z\\\",\\\"message\\\":\\\": current time 2025-10-07T13:46:36Z is after 2025-08-24T17:21:41Z]\\\\nI1007 13:46:36.558255 6945 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-diagnostics/network-check-target]} name:Service_openshift-network-diagnostics/network-check-target_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.219:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7594bb65-e742-44b3-a975-d639b1128be5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1007 13:46:36.558260 6945 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-apiserver-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"8b82f026-5975-4a1b-bb18-08d5d51147ec\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:37Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.243068 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e325099f-e0ff-433d-bef7-6909c5c86bea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecb7da009b7ed9e5c0412125be08cf486bf4f0e74480f2843877b0816944fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6dab7bc77a409dca97cffd302d0086021bf51146e8ec69a53cbddb87a4026aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b500df391c3b1ef8f9689a04e58c19e603667507a68fb8d65ea768690425085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:37Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.258861 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.258930 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.258954 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.258985 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.259009 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:37Z","lastTransitionTime":"2025-10-07T13:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.262519 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:37Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.281674 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:37Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.298287 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:37Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.317405 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:37Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.336134 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ab500b1b215518cbfc4602ee0671c6f3f68464467aa392ea5bc38ffb24bd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1454006ab0b9f83bd85e57a5e7389c81aa81ea5fb4d7c86a6315839de5c37812\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xpjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:37Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.353498 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zdps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zdps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:37Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.362385 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.362435 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.362451 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.362475 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.362494 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:37Z","lastTransitionTime":"2025-10-07T13:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.386710 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:37Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.410727 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:37Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.428915 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:37Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.450582 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:37Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.464878 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.464977 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.464997 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.465021 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.465040 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:37Z","lastTransitionTime":"2025-10-07T13:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.470854 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6898f5d8eb8c5a06853350d8189d1c393cef00392a0c6240b449dbb8617d9f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:46:31Z\\\",\\\"message\\\":\\\"2025-10-07T13:45:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_19b999ee-ab24-4a80-bd1e-5817a041f450\\\\n2025-10-07T13:45:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_19b999ee-ab24-4a80-bd1e-5817a041f450 to /host/opt/cni/bin/\\\\n2025-10-07T13:45:46Z [verbose] multus-daemon started\\\\n2025-10-07T13:45:46Z [verbose] Readiness Indicator file check\\\\n2025-10-07T13:46:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:46:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:37Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.489434 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1cbef019f0640b5d51b398d1fc7166c1e253fd33f78f0e50f9754bb1187a6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:37Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.505316 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:37Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.520813 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1ac79b-8fe3-40cd-925d-d840f0995abf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae966401c1e804da8649ca9c014cfc2eed8dbb12a0576cf90b6c7a3526513601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6d82be64b945bd47d6149c695f43152998644d31f0103bfff861199d5efb191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6d82be64b945bd47d6149c695f43152998644d31f0103bfff861199d5efb191\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:37Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.567693 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.567730 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.567739 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.567754 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.567765 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:37Z","lastTransitionTime":"2025-10-07T13:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.653031 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:37 crc kubenswrapper[4959]: E1007 13:46:37.653223 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.670070 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.670138 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.670150 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.670166 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.670179 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:37Z","lastTransitionTime":"2025-10-07T13:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.772875 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.772929 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.772945 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.772969 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.772988 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:37Z","lastTransitionTime":"2025-10-07T13:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.876082 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.876295 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.876320 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.876350 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.876372 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:37Z","lastTransitionTime":"2025-10-07T13:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.978888 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.978943 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.978952 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.978967 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:37 crc kubenswrapper[4959]: I1007 13:46:37.978980 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:37Z","lastTransitionTime":"2025-10-07T13:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.081365 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.081423 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.081441 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.081467 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.081485 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:38Z","lastTransitionTime":"2025-10-07T13:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.134277 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5sfv5_b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b/ovnkube-controller/3.log" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.139745 4959 scope.go:117] "RemoveContainer" containerID="fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69" Oct 07 13:46:38 crc kubenswrapper[4959]: E1007 13:46:38.140007 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5sfv5_openshift-ovn-kubernetes(b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.174907 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:46:36Z\\\",\\\"message\\\":\\\": current time 2025-10-07T13:46:36Z is after 2025-08-24T17:21:41Z]\\\\nI1007 13:46:36.558255 6945 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-diagnostics/network-check-target]} name:Service_openshift-network-diagnostics/network-check-target_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.219:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7594bb65-e742-44b3-a975-d639b1128be5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1007 13:46:36.558260 6945 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-apiserver-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"8b82f026-5975-4a1b-bb18-08d5d51147ec\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:46:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5sfv5_openshift-ovn-kubernetes(b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.184002 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.184039 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.184049 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.184066 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.184079 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:38Z","lastTransitionTime":"2025-10-07T13:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.194747 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e325099f-e0ff-433d-bef7-6909c5c86bea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecb7da009b7ed9e5c0412125be08cf486bf4f0e74480f2843877b0816944fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6dab7bc77a409dca97cffd302d0086021bf51146e8ec69a53cbddb87a4026aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b500df391c3b1ef8f9689a04e58c19e603667507a68fb8d65ea768690425085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.212667 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.231136 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.249322 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.264945 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.281122 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ab500b1b215518cbfc4602ee0671c6f3f68464467aa392ea5bc38ffb24bd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1454006ab0b9f83bd85e57a5e7389c81aa81ea5fb4d7c86a6315839de5c37812\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xpjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.285902 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.285949 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.285968 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.285995 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.286013 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:38Z","lastTransitionTime":"2025-10-07T13:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.296617 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zdps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zdps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.316829 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.331079 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.344285 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.354644 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.367484 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6898f5d8eb8c5a06853350d8189d1c393cef00392a0c6240b449dbb8617d9f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:46:31Z\\\",\\\"message\\\":\\\"2025-10-07T13:45:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_19b999ee-ab24-4a80-bd1e-5817a041f450\\\\n2025-10-07T13:45:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_19b999ee-ab24-4a80-bd1e-5817a041f450 to /host/opt/cni/bin/\\\\n2025-10-07T13:45:46Z [verbose] multus-daemon started\\\\n2025-10-07T13:45:46Z [verbose] Readiness Indicator file check\\\\n2025-10-07T13:46:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:46:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.384978 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1cbef019f0640b5d51b398d1fc7166c1e253fd33f78f0e50f9754bb1187a6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.388604 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.388642 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.388659 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.388680 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.388696 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:38Z","lastTransitionTime":"2025-10-07T13:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.419331 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.431388 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1ac79b-8fe3-40cd-925d-d840f0995abf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae966401c1e804da8649ca9c014cfc2eed8dbb12a0576cf90b6c7a3526513601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6d82be64b945bd47d6149c695f43152998644d31f0103bfff861199d5efb191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6d82be64b945bd47d6149c695f43152998644d31f0103bfff861199d5efb191\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.445400 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.457179 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.468870 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.491806 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.491881 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.491903 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.491931 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.491953 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:38Z","lastTransitionTime":"2025-10-07T13:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.595358 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.595411 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.595429 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.595456 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.595473 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:38Z","lastTransitionTime":"2025-10-07T13:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.653087 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.653194 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.653137 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:38 crc kubenswrapper[4959]: E1007 13:46:38.653597 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:38 crc kubenswrapper[4959]: E1007 13:46:38.653733 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:38 crc kubenswrapper[4959]: E1007 13:46:38.653808 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.674262 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e325099f-e0ff-433d-bef7-6909c5c86bea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecb7da009b7ed9e5c0412125be08cf486bf4f0e74480f2843877b0816944fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6dab7bc77a409dca97cffd302d0086021bf51146e8ec69a53cbddb87a4026aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b500df391c3b1ef8f9689a04e58c19e603667507a68fb8d65ea768690425085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.693429 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.698742 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.698795 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.698811 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.698834 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.698851 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:38Z","lastTransitionTime":"2025-10-07T13:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.712078 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.739369 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.769047 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:46:36Z\\\",\\\"message\\\":\\\": current time 2025-10-07T13:46:36Z is after 2025-08-24T17:21:41Z]\\\\nI1007 13:46:36.558255 6945 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-diagnostics/network-check-target]} name:Service_openshift-network-diagnostics/network-check-target_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.219:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7594bb65-e742-44b3-a975-d639b1128be5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1007 13:46:36.558260 6945 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-apiserver-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"8b82f026-5975-4a1b-bb18-08d5d51147ec\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:46:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5sfv5_openshift-ovn-kubernetes(b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.786927 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zdps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zdps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.801788 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.801827 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.801839 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.801856 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.801868 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:38Z","lastTransitionTime":"2025-10-07T13:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.817340 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.836460 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.853531 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.866640 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.881162 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.894934 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ab500b1b215518cbfc4602ee0671c6f3f68464467aa392ea5bc38ffb24bd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1454006ab0b9f83bd85e57a5e7389c81aa81ea5fb4d7c86a6315839de5c37812\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xpjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.903796 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.903834 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.903842 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.903856 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.903864 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:38Z","lastTransitionTime":"2025-10-07T13:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.907165 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.920555 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1ac79b-8fe3-40cd-925d-d840f0995abf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae966401c1e804da8649ca9c014cfc2eed8dbb12a0576cf90b6c7a3526513601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6d82be64b945bd47d6149c695f43152998644d31f0103bfff861199d5efb191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6d82be64b945bd47d6149c695f43152998644d31f0103bfff861199d5efb191\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.939289 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.956711 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.972215 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:38 crc kubenswrapper[4959]: I1007 13:46:38.989339 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6898f5d8eb8c5a06853350d8189d1c393cef00392a0c6240b449dbb8617d9f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:46:31Z\\\",\\\"message\\\":\\\"2025-10-07T13:45:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_19b999ee-ab24-4a80-bd1e-5817a041f450\\\\n2025-10-07T13:45:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_19b999ee-ab24-4a80-bd1e-5817a041f450 to /host/opt/cni/bin/\\\\n2025-10-07T13:45:46Z [verbose] multus-daemon started\\\\n2025-10-07T13:45:46Z [verbose] Readiness Indicator file check\\\\n2025-10-07T13:46:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:46:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:38Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.006561 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.006592 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.006600 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.006613 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.006622 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:39Z","lastTransitionTime":"2025-10-07T13:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.009450 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1cbef019f0640b5d51b398d1fc7166c1e253fd33f78f0e50f9754bb1187a6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:39Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.109575 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.109626 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.109641 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.109661 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.109675 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:39Z","lastTransitionTime":"2025-10-07T13:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.211676 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.211708 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.211718 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.211730 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.211739 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:39Z","lastTransitionTime":"2025-10-07T13:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.314226 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.314296 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.314323 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.314353 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.314370 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:39Z","lastTransitionTime":"2025-10-07T13:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.417532 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.417594 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.417615 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.417639 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.417656 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:39Z","lastTransitionTime":"2025-10-07T13:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.520860 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.521150 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.521287 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.521468 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.521597 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:39Z","lastTransitionTime":"2025-10-07T13:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.625033 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.625171 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.625191 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.625216 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.625233 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:39Z","lastTransitionTime":"2025-10-07T13:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.652586 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:39 crc kubenswrapper[4959]: E1007 13:46:39.652920 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.727570 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.727630 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.727646 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.727668 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.727685 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:39Z","lastTransitionTime":"2025-10-07T13:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.830218 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.830273 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.830290 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.830313 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.830330 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:39Z","lastTransitionTime":"2025-10-07T13:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.932441 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.932491 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.932506 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.932526 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:39 crc kubenswrapper[4959]: I1007 13:46:39.932541 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:39Z","lastTransitionTime":"2025-10-07T13:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.035637 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.035687 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.035698 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.035716 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.035728 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:40Z","lastTransitionTime":"2025-10-07T13:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.138580 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.138638 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.138656 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.138678 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.138695 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:40Z","lastTransitionTime":"2025-10-07T13:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.241573 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.241632 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.241649 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.241672 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.241693 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:40Z","lastTransitionTime":"2025-10-07T13:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.344690 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.344769 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.344794 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.344825 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.344848 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:40Z","lastTransitionTime":"2025-10-07T13:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.447949 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.447991 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.448004 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.448022 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.448036 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:40Z","lastTransitionTime":"2025-10-07T13:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.550878 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.550926 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.550942 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.550962 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.550979 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:40Z","lastTransitionTime":"2025-10-07T13:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.652567 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.652627 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:40 crc kubenswrapper[4959]: E1007 13:46:40.652669 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.653553 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.653611 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.653647 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.653680 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.653698 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:40Z","lastTransitionTime":"2025-10-07T13:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:40 crc kubenswrapper[4959]: E1007 13:46:40.653936 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.654007 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:40 crc kubenswrapper[4959]: E1007 13:46:40.654131 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.756611 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.756693 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.756712 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.756733 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.756787 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:40Z","lastTransitionTime":"2025-10-07T13:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.859052 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.859365 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.859432 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.859501 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.859560 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:40Z","lastTransitionTime":"2025-10-07T13:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.962980 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.963081 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.963144 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.963170 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:40 crc kubenswrapper[4959]: I1007 13:46:40.963186 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:40Z","lastTransitionTime":"2025-10-07T13:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.066431 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.066643 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.066707 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.066739 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.066806 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:41Z","lastTransitionTime":"2025-10-07T13:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.169595 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.169647 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.169665 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.169687 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.169705 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:41Z","lastTransitionTime":"2025-10-07T13:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.272757 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.272826 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.272841 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.272864 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.272879 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:41Z","lastTransitionTime":"2025-10-07T13:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.376086 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.376214 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.376237 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.376268 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.376289 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:41Z","lastTransitionTime":"2025-10-07T13:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.479752 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.479835 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.479854 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.479879 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.479897 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:41Z","lastTransitionTime":"2025-10-07T13:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.583042 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.583130 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.583149 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.583173 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.583190 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:41Z","lastTransitionTime":"2025-10-07T13:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.652851 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:41 crc kubenswrapper[4959]: E1007 13:46:41.653039 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.685914 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.685969 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.685987 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.686011 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.686029 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:41Z","lastTransitionTime":"2025-10-07T13:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.788820 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.789093 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.789161 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.789187 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.789205 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:41Z","lastTransitionTime":"2025-10-07T13:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.891461 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.891516 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.891532 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.891557 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.891575 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:41Z","lastTransitionTime":"2025-10-07T13:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.994781 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.994854 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.994872 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.994897 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:41 crc kubenswrapper[4959]: I1007 13:46:41.994916 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:41Z","lastTransitionTime":"2025-10-07T13:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.097358 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.097415 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.097431 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.097459 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.097477 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:42Z","lastTransitionTime":"2025-10-07T13:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.200969 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.201133 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.201166 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.201197 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.201221 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:42Z","lastTransitionTime":"2025-10-07T13:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.304636 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.304684 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.304696 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.304714 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.304728 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:42Z","lastTransitionTime":"2025-10-07T13:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.414263 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.414320 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.414338 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.414407 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.414433 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:42Z","lastTransitionTime":"2025-10-07T13:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.517634 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.517688 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.517703 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.517726 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.517742 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:42Z","lastTransitionTime":"2025-10-07T13:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.620308 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.620356 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.620370 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.620390 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.620404 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:42Z","lastTransitionTime":"2025-10-07T13:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.653237 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.653332 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:42 crc kubenswrapper[4959]: E1007 13:46:42.653384 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.653424 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:42 crc kubenswrapper[4959]: E1007 13:46:42.653579 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:42 crc kubenswrapper[4959]: E1007 13:46:42.653725 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.669441 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.669666 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:42 crc kubenswrapper[4959]: E1007 13:46:42.669682 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:46.669653403 +0000 UTC m=+148.753057818 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.669750 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.669827 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:42 crc kubenswrapper[4959]: E1007 13:46:42.669852 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.669875 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:42 crc kubenswrapper[4959]: E1007 13:46:42.669889 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 13:46:42 crc kubenswrapper[4959]: E1007 13:46:42.669909 4959 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:46:42 crc kubenswrapper[4959]: E1007 13:46:42.669953 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 07 13:46:42 crc kubenswrapper[4959]: E1007 13:46:42.669977 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 07 13:46:42 crc kubenswrapper[4959]: E1007 13:46:42.669980 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-07 13:47:46.669955111 +0000 UTC m=+148.753359466 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:46:42 crc kubenswrapper[4959]: E1007 13:46:42.669992 4959 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:46:42 crc kubenswrapper[4959]: E1007 13:46:42.670024 4959 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 13:46:42 crc kubenswrapper[4959]: E1007 13:46:42.670038 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-07 13:47:46.670027303 +0000 UTC m=+148.753431738 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 07 13:46:42 crc kubenswrapper[4959]: E1007 13:46:42.670091 4959 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 13:46:42 crc kubenswrapper[4959]: E1007 13:46:42.670094 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 13:47:46.670072094 +0000 UTC m=+148.753476459 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 07 13:46:42 crc kubenswrapper[4959]: E1007 13:46:42.670158 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-07 13:47:46.670147896 +0000 UTC m=+148.753552391 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.723659 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.723750 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.723770 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.723825 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.723842 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:42Z","lastTransitionTime":"2025-10-07T13:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.826347 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.826415 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.826433 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.826457 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.826474 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:42Z","lastTransitionTime":"2025-10-07T13:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.930183 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.930271 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.930289 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.930341 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:42 crc kubenswrapper[4959]: I1007 13:46:42.930360 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:42Z","lastTransitionTime":"2025-10-07T13:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.034151 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.034257 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.034278 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.034303 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.034363 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:43Z","lastTransitionTime":"2025-10-07T13:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.137407 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.137460 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.137477 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.137500 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.137521 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:43Z","lastTransitionTime":"2025-10-07T13:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.240378 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.240564 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.240595 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.240623 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.240644 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:43Z","lastTransitionTime":"2025-10-07T13:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.343933 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.344012 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.344040 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.344071 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.344091 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:43Z","lastTransitionTime":"2025-10-07T13:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.447686 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.447776 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.447794 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.447850 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.447868 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:43Z","lastTransitionTime":"2025-10-07T13:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.551061 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.551309 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.551350 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.551378 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.551399 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:43Z","lastTransitionTime":"2025-10-07T13:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.652882 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:43 crc kubenswrapper[4959]: E1007 13:46:43.653526 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.654273 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.654339 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.654363 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.654389 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.654411 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:43Z","lastTransitionTime":"2025-10-07T13:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.757807 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.757945 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.757967 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.757992 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.758009 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:43Z","lastTransitionTime":"2025-10-07T13:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.860568 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.860603 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.860614 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.860627 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.860635 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:43Z","lastTransitionTime":"2025-10-07T13:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.963027 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.963268 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.963291 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.963315 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:43 crc kubenswrapper[4959]: I1007 13:46:43.963331 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:43Z","lastTransitionTime":"2025-10-07T13:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.066623 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.066719 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.066736 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.066762 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.066778 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:44Z","lastTransitionTime":"2025-10-07T13:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.170210 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.170281 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.170298 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.170322 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.170342 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:44Z","lastTransitionTime":"2025-10-07T13:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.273298 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.273379 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.273402 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.273429 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.273446 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:44Z","lastTransitionTime":"2025-10-07T13:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.376751 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.376823 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.376845 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.376876 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.376897 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:44Z","lastTransitionTime":"2025-10-07T13:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.479316 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.479347 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.479354 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.479366 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.479374 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:44Z","lastTransitionTime":"2025-10-07T13:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.581964 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.582083 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.582138 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.582165 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.582182 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:44Z","lastTransitionTime":"2025-10-07T13:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.653186 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.653231 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.653229 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:44 crc kubenswrapper[4959]: E1007 13:46:44.653376 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:44 crc kubenswrapper[4959]: E1007 13:46:44.653494 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:44 crc kubenswrapper[4959]: E1007 13:46:44.653677 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.684605 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.684670 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.684694 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.684722 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.684744 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:44Z","lastTransitionTime":"2025-10-07T13:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.787866 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.787930 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.787948 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.787973 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.787992 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:44Z","lastTransitionTime":"2025-10-07T13:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.890624 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.890708 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.890732 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.890764 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.890787 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:44Z","lastTransitionTime":"2025-10-07T13:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.993744 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.993821 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.993851 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.993884 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:44 crc kubenswrapper[4959]: I1007 13:46:44.993905 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:44Z","lastTransitionTime":"2025-10-07T13:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.097928 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.097997 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.098019 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.098046 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.098065 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:45Z","lastTransitionTime":"2025-10-07T13:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.200665 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.201003 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.201236 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.201421 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.201561 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:45Z","lastTransitionTime":"2025-10-07T13:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.305279 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.305660 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.305678 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.305702 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.305718 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:45Z","lastTransitionTime":"2025-10-07T13:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.408615 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.408675 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.408693 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.408717 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.408735 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:45Z","lastTransitionTime":"2025-10-07T13:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.511424 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.511484 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.511501 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.511524 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.511544 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:45Z","lastTransitionTime":"2025-10-07T13:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.615235 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.615290 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.615309 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.615332 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.615349 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:45Z","lastTransitionTime":"2025-10-07T13:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.653066 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:45 crc kubenswrapper[4959]: E1007 13:46:45.653440 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.717894 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.717952 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.717969 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.717993 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.718011 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:45Z","lastTransitionTime":"2025-10-07T13:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.788881 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.789189 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.789370 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.789519 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.789655 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:45Z","lastTransitionTime":"2025-10-07T13:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:45 crc kubenswrapper[4959]: E1007 13:46:45.810981 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.816226 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.816430 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.816594 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.816728 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.816873 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:45Z","lastTransitionTime":"2025-10-07T13:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:45 crc kubenswrapper[4959]: E1007 13:46:45.839504 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.845092 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.845178 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.845194 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.845219 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.845237 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:45Z","lastTransitionTime":"2025-10-07T13:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:45 crc kubenswrapper[4959]: E1007 13:46:45.865913 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.871647 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.871883 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.872024 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.872196 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.872374 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:45Z","lastTransitionTime":"2025-10-07T13:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:45 crc kubenswrapper[4959]: E1007 13:46:45.892259 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.897357 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.897421 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.897438 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.897463 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.897480 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:45Z","lastTransitionTime":"2025-10-07T13:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:45 crc kubenswrapper[4959]: E1007 13:46:45.919053 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:45Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:45 crc kubenswrapper[4959]: E1007 13:46:45.919715 4959 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.922745 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.922793 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.922819 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.922848 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:45 crc kubenswrapper[4959]: I1007 13:46:45.922870 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:45Z","lastTransitionTime":"2025-10-07T13:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.025841 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.026239 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.026430 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.026576 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.026717 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:46Z","lastTransitionTime":"2025-10-07T13:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.130204 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.130264 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.130282 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.130310 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.130333 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:46Z","lastTransitionTime":"2025-10-07T13:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.233010 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.233077 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.233094 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.233144 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.233161 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:46Z","lastTransitionTime":"2025-10-07T13:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.336541 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.336657 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.336682 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.336713 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.336734 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:46Z","lastTransitionTime":"2025-10-07T13:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.440448 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.440515 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.440538 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.440572 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.440596 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:46Z","lastTransitionTime":"2025-10-07T13:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.543532 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.543609 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.543632 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.543660 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.543678 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:46Z","lastTransitionTime":"2025-10-07T13:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.646452 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.646509 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.646526 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.646550 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.646568 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:46Z","lastTransitionTime":"2025-10-07T13:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.652806 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.652917 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:46 crc kubenswrapper[4959]: E1007 13:46:46.653010 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.653069 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:46 crc kubenswrapper[4959]: E1007 13:46:46.653278 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:46 crc kubenswrapper[4959]: E1007 13:46:46.653388 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.750292 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.750872 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.750904 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.750933 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.750963 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:46Z","lastTransitionTime":"2025-10-07T13:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.855700 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.855765 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.855784 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.855826 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.855845 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:46Z","lastTransitionTime":"2025-10-07T13:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.959024 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.959083 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.959121 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.959144 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:46 crc kubenswrapper[4959]: I1007 13:46:46.959160 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:46Z","lastTransitionTime":"2025-10-07T13:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.062392 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.062455 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.062473 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.062500 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.062517 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:47Z","lastTransitionTime":"2025-10-07T13:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.167151 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.167233 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.167255 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.167278 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.167296 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:47Z","lastTransitionTime":"2025-10-07T13:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.270316 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.270386 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.270411 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.270439 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.270460 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:47Z","lastTransitionTime":"2025-10-07T13:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.373878 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.373949 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.373968 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.373993 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.374011 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:47Z","lastTransitionTime":"2025-10-07T13:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.477645 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.477710 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.477727 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.477751 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.477768 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:47Z","lastTransitionTime":"2025-10-07T13:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.580373 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.580467 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.580505 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.580538 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.580564 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:47Z","lastTransitionTime":"2025-10-07T13:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.653437 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:47 crc kubenswrapper[4959]: E1007 13:46:47.653678 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.683515 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.683576 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.683592 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.683617 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.683633 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:47Z","lastTransitionTime":"2025-10-07T13:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.786685 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.786738 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.786754 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.786779 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.786798 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:47Z","lastTransitionTime":"2025-10-07T13:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.889749 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.889815 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.889840 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.889871 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.889898 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:47Z","lastTransitionTime":"2025-10-07T13:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.992634 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.992691 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.992709 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.992735 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:47 crc kubenswrapper[4959]: I1007 13:46:47.992753 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:47Z","lastTransitionTime":"2025-10-07T13:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.095083 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.095134 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.095143 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.095157 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.095166 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:48Z","lastTransitionTime":"2025-10-07T13:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.197632 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.197673 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.197685 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.197701 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.197711 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:48Z","lastTransitionTime":"2025-10-07T13:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.301334 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.301409 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.301462 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.301491 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.301510 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:48Z","lastTransitionTime":"2025-10-07T13:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.403964 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.404016 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.404028 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.404045 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.404057 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:48Z","lastTransitionTime":"2025-10-07T13:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.506666 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.506729 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.506748 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.506774 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.506793 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:48Z","lastTransitionTime":"2025-10-07T13:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.610000 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.610041 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.610049 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.610064 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.610075 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:48Z","lastTransitionTime":"2025-10-07T13:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.653335 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.653396 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:48 crc kubenswrapper[4959]: E1007 13:46:48.653491 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.653523 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:48 crc kubenswrapper[4959]: E1007 13:46:48.653690 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:48 crc kubenswrapper[4959]: E1007 13:46:48.653835 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.672268 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.692219 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.708424 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.713069 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.713161 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.713181 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.713207 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.713230 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:48Z","lastTransitionTime":"2025-10-07T13:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.739532 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:46:36Z\\\",\\\"message\\\":\\\": current time 2025-10-07T13:46:36Z is after 2025-08-24T17:21:41Z]\\\\nI1007 13:46:36.558255 6945 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-diagnostics/network-check-target]} name:Service_openshift-network-diagnostics/network-check-target_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.219:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7594bb65-e742-44b3-a975-d639b1128be5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1007 13:46:36.558260 6945 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-apiserver-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"8b82f026-5975-4a1b-bb18-08d5d51147ec\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:46:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5sfv5_openshift-ovn-kubernetes(b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.759234 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e325099f-e0ff-433d-bef7-6909c5c86bea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecb7da009b7ed9e5c0412125be08cf486bf4f0e74480f2843877b0816944fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6dab7bc77a409dca97cffd302d0086021bf51146e8ec69a53cbddb87a4026aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b500df391c3b1ef8f9689a04e58c19e603667507a68fb8d65ea768690425085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.784467 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.805055 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.816082 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.816150 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.816163 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.816181 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.816193 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:48Z","lastTransitionTime":"2025-10-07T13:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.816615 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.833055 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.846722 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ab500b1b215518cbfc4602ee0671c6f3f68464467aa392ea5bc38ffb24bd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1454006ab0b9f83bd85e57a5e7389c81aa81ea5fb4d7c86a6315839de5c37812\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xpjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.858478 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zdps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zdps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.882016 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.906427 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.919941 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.920160 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.920276 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.920382 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.920480 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:48Z","lastTransitionTime":"2025-10-07T13:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.925122 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.941948 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.959179 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6898f5d8eb8c5a06853350d8189d1c393cef00392a0c6240b449dbb8617d9f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:46:31Z\\\",\\\"message\\\":\\\"2025-10-07T13:45:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_19b999ee-ab24-4a80-bd1e-5817a041f450\\\\n2025-10-07T13:45:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_19b999ee-ab24-4a80-bd1e-5817a041f450 to /host/opt/cni/bin/\\\\n2025-10-07T13:45:46Z [verbose] multus-daemon started\\\\n2025-10-07T13:45:46Z [verbose] Readiness Indicator file check\\\\n2025-10-07T13:46:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:46:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.978288 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1cbef019f0640b5d51b398d1fc7166c1e253fd33f78f0e50f9754bb1187a6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:48 crc kubenswrapper[4959]: I1007 13:46:48.993739 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:48Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.009556 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1ac79b-8fe3-40cd-925d-d840f0995abf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae966401c1e804da8649ca9c014cfc2eed8dbb12a0576cf90b6c7a3526513601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6d82be64b945bd47d6149c695f43152998644d31f0103bfff861199d5efb191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6d82be64b945bd47d6149c695f43152998644d31f0103bfff861199d5efb191\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:49Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.024187 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.024253 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.024272 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.024300 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.024322 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:49Z","lastTransitionTime":"2025-10-07T13:46:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.127628 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.127948 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.128052 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.128189 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.128301 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:49Z","lastTransitionTime":"2025-10-07T13:46:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.230739 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.230786 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.230799 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.230816 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.230829 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:49Z","lastTransitionTime":"2025-10-07T13:46:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.332987 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.333034 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.333051 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.333073 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.333090 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:49Z","lastTransitionTime":"2025-10-07T13:46:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.435945 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.436020 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.436045 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.436073 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.436095 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:49Z","lastTransitionTime":"2025-10-07T13:46:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.539445 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.539757 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.539780 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.539810 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.539836 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:49Z","lastTransitionTime":"2025-10-07T13:46:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.642957 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.643025 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.643048 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.643077 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.643095 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:49Z","lastTransitionTime":"2025-10-07T13:46:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.652872 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:49 crc kubenswrapper[4959]: E1007 13:46:49.653206 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.746948 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.747007 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.747025 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.747046 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.747063 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:49Z","lastTransitionTime":"2025-10-07T13:46:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.850245 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.850312 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.850330 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.850353 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.850371 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:49Z","lastTransitionTime":"2025-10-07T13:46:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.954066 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.954150 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.954169 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.954193 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:49 crc kubenswrapper[4959]: I1007 13:46:49.954211 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:49Z","lastTransitionTime":"2025-10-07T13:46:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.056870 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.056944 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.056968 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.056999 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.057024 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:50Z","lastTransitionTime":"2025-10-07T13:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.162555 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.162606 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.162624 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.162645 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.162662 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:50Z","lastTransitionTime":"2025-10-07T13:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.266146 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.266202 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.266219 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.266245 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.266265 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:50Z","lastTransitionTime":"2025-10-07T13:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.370446 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.370529 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.370551 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.370582 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.370603 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:50Z","lastTransitionTime":"2025-10-07T13:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.473177 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.473250 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.473277 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.473309 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.473335 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:50Z","lastTransitionTime":"2025-10-07T13:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.577332 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.577401 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.577418 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.577444 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.577465 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:50Z","lastTransitionTime":"2025-10-07T13:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.653021 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.653388 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:50 crc kubenswrapper[4959]: E1007 13:46:50.653383 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.653461 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:50 crc kubenswrapper[4959]: E1007 13:46:50.653755 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:50 crc kubenswrapper[4959]: E1007 13:46:50.654279 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.680860 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.680919 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.680939 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.680966 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.680985 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:50Z","lastTransitionTime":"2025-10-07T13:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.783579 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.783648 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.783668 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.783704 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.783724 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:50Z","lastTransitionTime":"2025-10-07T13:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.886967 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.887049 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.887077 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.887142 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.887167 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:50Z","lastTransitionTime":"2025-10-07T13:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.990848 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.990907 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.990925 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.990948 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:50 crc kubenswrapper[4959]: I1007 13:46:50.990970 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:50Z","lastTransitionTime":"2025-10-07T13:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.093643 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.093691 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.093703 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.093721 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.093735 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:51Z","lastTransitionTime":"2025-10-07T13:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.197009 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.197080 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.197138 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.197170 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.197192 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:51Z","lastTransitionTime":"2025-10-07T13:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.301316 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.301398 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.301427 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.301459 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.301482 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:51Z","lastTransitionTime":"2025-10-07T13:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.404364 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.404418 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.404466 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.404495 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.404515 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:51Z","lastTransitionTime":"2025-10-07T13:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.507204 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.507310 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.507329 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.507356 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.507376 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:51Z","lastTransitionTime":"2025-10-07T13:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.610943 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.611006 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.611023 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.611046 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.611063 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:51Z","lastTransitionTime":"2025-10-07T13:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.652912 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:51 crc kubenswrapper[4959]: E1007 13:46:51.653304 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.715777 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.715854 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.715879 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.715909 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.715948 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:51Z","lastTransitionTime":"2025-10-07T13:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.819144 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.819235 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.819252 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.819282 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.819306 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:51Z","lastTransitionTime":"2025-10-07T13:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.922359 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.922421 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.922438 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.922462 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:51 crc kubenswrapper[4959]: I1007 13:46:51.922479 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:51Z","lastTransitionTime":"2025-10-07T13:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.025155 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.025201 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.025220 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.025244 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.025260 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:52Z","lastTransitionTime":"2025-10-07T13:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.128013 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.128313 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.128387 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.128418 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.128435 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:52Z","lastTransitionTime":"2025-10-07T13:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.231573 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.231641 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.231658 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.231686 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.231714 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:52Z","lastTransitionTime":"2025-10-07T13:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.333981 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.334014 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.334022 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.334034 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.334043 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:52Z","lastTransitionTime":"2025-10-07T13:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.436507 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.436561 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.436576 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.436595 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.436609 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:52Z","lastTransitionTime":"2025-10-07T13:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.540061 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.540181 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.540200 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.540237 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.540255 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:52Z","lastTransitionTime":"2025-10-07T13:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.643810 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.643889 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.643913 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.643943 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.643966 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:52Z","lastTransitionTime":"2025-10-07T13:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.653235 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.653291 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.653380 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:52 crc kubenswrapper[4959]: E1007 13:46:52.653589 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:52 crc kubenswrapper[4959]: E1007 13:46:52.653874 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:52 crc kubenswrapper[4959]: E1007 13:46:52.654529 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.655019 4959 scope.go:117] "RemoveContainer" containerID="fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69" Oct 07 13:46:52 crc kubenswrapper[4959]: E1007 13:46:52.655300 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5sfv5_openshift-ovn-kubernetes(b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.747655 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.747717 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.747739 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.747767 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.747789 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:52Z","lastTransitionTime":"2025-10-07T13:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.851026 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.851091 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.851143 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.851175 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.851196 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:52Z","lastTransitionTime":"2025-10-07T13:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.954363 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.954426 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.954443 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.954467 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:52 crc kubenswrapper[4959]: I1007 13:46:52.954530 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:52Z","lastTransitionTime":"2025-10-07T13:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.057690 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.057763 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.057786 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.057816 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.057837 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:53Z","lastTransitionTime":"2025-10-07T13:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.160750 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.160813 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.160836 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.160866 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.160888 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:53Z","lastTransitionTime":"2025-10-07T13:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.263024 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.263087 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.263129 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.263154 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.263172 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:53Z","lastTransitionTime":"2025-10-07T13:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.366376 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.366454 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.366472 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.366501 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.366523 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:53Z","lastTransitionTime":"2025-10-07T13:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.470079 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.470178 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.470197 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.470223 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.470246 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:53Z","lastTransitionTime":"2025-10-07T13:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.573405 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.573492 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.573515 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.573549 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.573573 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:53Z","lastTransitionTime":"2025-10-07T13:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.653461 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:53 crc kubenswrapper[4959]: E1007 13:46:53.653669 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.677672 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.677726 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.677743 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.677761 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.677774 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:53Z","lastTransitionTime":"2025-10-07T13:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.781708 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.781850 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.781873 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.781902 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.781923 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:53Z","lastTransitionTime":"2025-10-07T13:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.885789 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.885853 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.885881 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.885912 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.885937 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:53Z","lastTransitionTime":"2025-10-07T13:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.989062 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.989173 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.989201 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.989231 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:53 crc kubenswrapper[4959]: I1007 13:46:53.989255 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:53Z","lastTransitionTime":"2025-10-07T13:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.092969 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.093029 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.093043 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.093070 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.093086 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:54Z","lastTransitionTime":"2025-10-07T13:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.196922 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.196985 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.197009 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.197037 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.197060 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:54Z","lastTransitionTime":"2025-10-07T13:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.299882 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.299923 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.299939 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.299963 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.299980 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:54Z","lastTransitionTime":"2025-10-07T13:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.403149 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.403199 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.403215 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.403235 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.403252 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:54Z","lastTransitionTime":"2025-10-07T13:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.506180 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.506242 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.506265 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.506291 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.506315 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:54Z","lastTransitionTime":"2025-10-07T13:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.609876 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.609943 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.609965 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.609997 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.610017 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:54Z","lastTransitionTime":"2025-10-07T13:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.652520 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.652646 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.652724 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:54 crc kubenswrapper[4959]: E1007 13:46:54.652665 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:54 crc kubenswrapper[4959]: E1007 13:46:54.652876 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:54 crc kubenswrapper[4959]: E1007 13:46:54.652989 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.712049 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.712130 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.712149 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.712170 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.712186 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:54Z","lastTransitionTime":"2025-10-07T13:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.814891 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.814938 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.814947 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.814960 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.814970 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:54Z","lastTransitionTime":"2025-10-07T13:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.918342 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.918409 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.918426 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.918452 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:54 crc kubenswrapper[4959]: I1007 13:46:54.918472 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:54Z","lastTransitionTime":"2025-10-07T13:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.022927 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.022996 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.023015 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.023038 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.023056 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:55Z","lastTransitionTime":"2025-10-07T13:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.126670 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.126760 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.126812 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.126836 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.126852 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:55Z","lastTransitionTime":"2025-10-07T13:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.229867 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.229947 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.229965 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.229993 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.230012 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:55Z","lastTransitionTime":"2025-10-07T13:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.332773 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.332850 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.332878 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.332906 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.332931 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:55Z","lastTransitionTime":"2025-10-07T13:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.436030 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.436073 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.436087 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.436123 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.436135 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:55Z","lastTransitionTime":"2025-10-07T13:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.539450 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.539507 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.539523 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.539542 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.539555 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:55Z","lastTransitionTime":"2025-10-07T13:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.641934 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.642000 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.642027 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.642058 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.642078 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:55Z","lastTransitionTime":"2025-10-07T13:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.653513 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:55 crc kubenswrapper[4959]: E1007 13:46:55.653688 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.744922 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.744988 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.745012 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.745042 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.745063 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:55Z","lastTransitionTime":"2025-10-07T13:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.847806 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.847865 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.847883 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.847907 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.847925 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:55Z","lastTransitionTime":"2025-10-07T13:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.950903 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.950956 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.950969 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.950985 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:55 crc kubenswrapper[4959]: I1007 13:46:55.950996 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:55Z","lastTransitionTime":"2025-10-07T13:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.054189 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.054230 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.054242 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.054257 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.054266 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:56Z","lastTransitionTime":"2025-10-07T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.097797 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.097833 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.097848 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.097862 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.097871 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:56Z","lastTransitionTime":"2025-10-07T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:56 crc kubenswrapper[4959]: E1007 13:46:56.117495 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:56Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.122636 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.122740 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.122809 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.122842 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.122865 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:56Z","lastTransitionTime":"2025-10-07T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:56 crc kubenswrapper[4959]: E1007 13:46:56.143309 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:56Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.148561 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.148613 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.148639 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.148667 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.148689 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:56Z","lastTransitionTime":"2025-10-07T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:56 crc kubenswrapper[4959]: E1007 13:46:56.168561 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:56Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.173586 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.173646 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.173669 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.173695 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.173716 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:56Z","lastTransitionTime":"2025-10-07T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:56 crc kubenswrapper[4959]: E1007 13:46:56.193959 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:56Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.199275 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.199329 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.199341 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.199356 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.199371 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:56Z","lastTransitionTime":"2025-10-07T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:56 crc kubenswrapper[4959]: E1007 13:46:56.219674 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:56Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:56 crc kubenswrapper[4959]: E1007 13:46:56.219986 4959 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.222297 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.222353 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.222371 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.222393 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.222411 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:56Z","lastTransitionTime":"2025-10-07T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.325514 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.325572 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.325589 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.325611 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.325628 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:56Z","lastTransitionTime":"2025-10-07T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.429042 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.429135 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.429154 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.429178 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.429195 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:56Z","lastTransitionTime":"2025-10-07T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.531410 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.531461 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.531474 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.531490 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.531503 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:56Z","lastTransitionTime":"2025-10-07T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.634845 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.634911 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.634935 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.634965 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.634986 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:56Z","lastTransitionTime":"2025-10-07T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.652626 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:56 crc kubenswrapper[4959]: E1007 13:46:56.652821 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.652914 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.652930 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:56 crc kubenswrapper[4959]: E1007 13:46:56.653221 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:56 crc kubenswrapper[4959]: E1007 13:46:56.653367 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.738131 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.738183 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.738200 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.738226 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.738242 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:56Z","lastTransitionTime":"2025-10-07T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.841699 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.841748 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.841765 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.841787 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.841805 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:56Z","lastTransitionTime":"2025-10-07T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.945329 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.945393 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.945423 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.945469 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:56 crc kubenswrapper[4959]: I1007 13:46:56.945495 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:56Z","lastTransitionTime":"2025-10-07T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.048838 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.048925 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.048951 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.048980 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.048997 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:57Z","lastTransitionTime":"2025-10-07T13:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.151593 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.151645 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.151664 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.151691 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.151710 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:57Z","lastTransitionTime":"2025-10-07T13:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.254087 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.254200 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.254219 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.254258 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.254278 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:57Z","lastTransitionTime":"2025-10-07T13:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.357668 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.357737 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.357756 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.357788 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.357808 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:57Z","lastTransitionTime":"2025-10-07T13:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.460059 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.460159 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.460190 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.460218 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.460239 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:57Z","lastTransitionTime":"2025-10-07T13:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.562932 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.562986 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.563002 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.563027 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.563044 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:57Z","lastTransitionTime":"2025-10-07T13:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.653305 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:57 crc kubenswrapper[4959]: E1007 13:46:57.653511 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.665530 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.665604 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.665623 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.665651 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.665669 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:57Z","lastTransitionTime":"2025-10-07T13:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.768666 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.768735 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.768757 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.768787 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.768809 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:57Z","lastTransitionTime":"2025-10-07T13:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.871960 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.872017 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.872034 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.872056 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.872078 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:57Z","lastTransitionTime":"2025-10-07T13:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.974752 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.974822 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.974845 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.974871 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:57 crc kubenswrapper[4959]: I1007 13:46:57.974888 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:57Z","lastTransitionTime":"2025-10-07T13:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.078009 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.078060 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.078083 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.078150 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.078176 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:58Z","lastTransitionTime":"2025-10-07T13:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.181472 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.181532 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.181553 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.181580 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.181602 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:58Z","lastTransitionTime":"2025-10-07T13:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.284327 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.284397 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.284420 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.284447 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.284470 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:58Z","lastTransitionTime":"2025-10-07T13:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.387598 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.387667 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.387690 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.387720 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.387743 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:58Z","lastTransitionTime":"2025-10-07T13:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.490666 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.490734 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.490753 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.490777 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.490800 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:58Z","lastTransitionTime":"2025-10-07T13:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.601261 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.601347 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.601367 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.601394 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.601417 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:58Z","lastTransitionTime":"2025-10-07T13:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.652864 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.652966 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.652888 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:46:58 crc kubenswrapper[4959]: E1007 13:46:58.653279 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:46:58 crc kubenswrapper[4959]: E1007 13:46:58.653444 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:46:58 crc kubenswrapper[4959]: E1007 13:46:58.654238 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.682927 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1cbef019f0640b5d51b398d1fc7166c1e253fd33f78f0e50f9754bb1187a6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.700591 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.704794 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.704825 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.704833 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.704846 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.704856 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:58Z","lastTransitionTime":"2025-10-07T13:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.716005 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1ac79b-8fe3-40cd-925d-d840f0995abf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae966401c1e804da8649ca9c014cfc2eed8dbb12a0576cf90b6c7a3526513601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6d82be64b945bd47d6149c695f43152998644d31f0103bfff861199d5efb191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6d82be64b945bd47d6149c695f43152998644d31f0103bfff861199d5efb191\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.741762 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.761114 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.779188 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.796914 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6898f5d8eb8c5a06853350d8189d1c393cef00392a0c6240b449dbb8617d9f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:46:31Z\\\",\\\"message\\\":\\\"2025-10-07T13:45:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_19b999ee-ab24-4a80-bd1e-5817a041f450\\\\n2025-10-07T13:45:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_19b999ee-ab24-4a80-bd1e-5817a041f450 to /host/opt/cni/bin/\\\\n2025-10-07T13:45:46Z [verbose] multus-daemon started\\\\n2025-10-07T13:45:46Z [verbose] Readiness Indicator file check\\\\n2025-10-07T13:46:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:46:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.807700 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.807762 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.807779 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.807801 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.807818 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:58Z","lastTransitionTime":"2025-10-07T13:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.815074 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e325099f-e0ff-433d-bef7-6909c5c86bea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecb7da009b7ed9e5c0412125be08cf486bf4f0e74480f2843877b0816944fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6dab7bc77a409dca97cffd302d0086021bf51146e8ec69a53cbddb87a4026aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b500df391c3b1ef8f9689a04e58c19e603667507a68fb8d65ea768690425085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.833334 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.850391 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.870543 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.900384 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:46:36Z\\\",\\\"message\\\":\\\": current time 2025-10-07T13:46:36Z is after 2025-08-24T17:21:41Z]\\\\nI1007 13:46:36.558255 6945 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-diagnostics/network-check-target]} name:Service_openshift-network-diagnostics/network-check-target_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.219:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7594bb65-e742-44b3-a975-d639b1128be5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1007 13:46:36.558260 6945 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-apiserver-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"8b82f026-5975-4a1b-bb18-08d5d51147ec\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:46:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5sfv5_openshift-ovn-kubernetes(b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.910506 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.910580 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.910606 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.910638 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.910662 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:58Z","lastTransitionTime":"2025-10-07T13:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.921049 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ab500b1b215518cbfc4602ee0671c6f3f68464467aa392ea5bc38ffb24bd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1454006ab0b9f83bd85e57a5e7389c81aa81ea5fb4d7c86a6315839de5c37812\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xpjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.939336 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zdps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zdps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.972138 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:58 crc kubenswrapper[4959]: I1007 13:46:58.993865 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:58Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.014587 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.014656 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.014679 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.014709 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.014730 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:59Z","lastTransitionTime":"2025-10-07T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.016600 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:59Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.032035 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:59Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.048965 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:46:59Z is after 2025-08-24T17:21:41Z" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.117959 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.118020 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.118040 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.118064 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.118130 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:59Z","lastTransitionTime":"2025-10-07T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.220680 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.220754 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.220777 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.220805 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.220828 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:59Z","lastTransitionTime":"2025-10-07T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.322893 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.322935 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.322947 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.322964 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.322976 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:59Z","lastTransitionTime":"2025-10-07T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.426463 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.426568 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.427059 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.427247 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.427588 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:59Z","lastTransitionTime":"2025-10-07T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.534536 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.534618 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.534641 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.534668 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.534706 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:59Z","lastTransitionTime":"2025-10-07T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.637637 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.637689 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.637703 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.637720 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.637732 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:59Z","lastTransitionTime":"2025-10-07T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.652916 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:46:59 crc kubenswrapper[4959]: E1007 13:46:59.653031 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.740684 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.740727 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.740742 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.740762 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.740775 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:59Z","lastTransitionTime":"2025-10-07T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.844064 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.844160 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.844177 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.844201 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.844219 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:59Z","lastTransitionTime":"2025-10-07T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.947167 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.947205 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.947215 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.947232 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:46:59 crc kubenswrapper[4959]: I1007 13:46:59.947244 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:46:59Z","lastTransitionTime":"2025-10-07T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.049377 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.049443 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.049467 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.049491 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.049508 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:00Z","lastTransitionTime":"2025-10-07T13:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.152869 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.152949 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.152977 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.153007 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.153028 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:00Z","lastTransitionTime":"2025-10-07T13:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.256063 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.256183 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.256210 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.256232 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.256249 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:00Z","lastTransitionTime":"2025-10-07T13:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.359791 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.359839 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.359854 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.359871 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.359884 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:00Z","lastTransitionTime":"2025-10-07T13:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.463925 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.463999 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.464021 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.464048 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.464069 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:00Z","lastTransitionTime":"2025-10-07T13:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.566588 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.566652 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.566671 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.566694 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.566712 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:00Z","lastTransitionTime":"2025-10-07T13:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.653363 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.653473 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:47:00 crc kubenswrapper[4959]: E1007 13:47:00.653630 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.653702 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:47:00 crc kubenswrapper[4959]: E1007 13:47:00.653844 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:47:00 crc kubenswrapper[4959]: E1007 13:47:00.654032 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.669728 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.669773 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.669791 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.669812 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.669828 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:00Z","lastTransitionTime":"2025-10-07T13:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.772432 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.772496 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.772516 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.772539 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.772556 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:00Z","lastTransitionTime":"2025-10-07T13:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.875877 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.875932 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.875950 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.875973 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.875994 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:00Z","lastTransitionTime":"2025-10-07T13:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.979525 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.979572 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.979587 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.979604 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:00 crc kubenswrapper[4959]: I1007 13:47:00.979616 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:00Z","lastTransitionTime":"2025-10-07T13:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.082433 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.082528 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.082548 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.082573 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.082590 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:01Z","lastTransitionTime":"2025-10-07T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.184686 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.184729 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.184741 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.184757 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.184771 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:01Z","lastTransitionTime":"2025-10-07T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.286855 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.286905 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.286917 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.286934 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.286945 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:01Z","lastTransitionTime":"2025-10-07T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.389990 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.390058 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.390075 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.390133 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.390153 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:01Z","lastTransitionTime":"2025-10-07T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.492222 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.492262 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.492274 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.492288 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.492298 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:01Z","lastTransitionTime":"2025-10-07T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.595489 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.595550 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.595567 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.595590 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.595606 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:01Z","lastTransitionTime":"2025-10-07T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.653251 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:47:01 crc kubenswrapper[4959]: E1007 13:47:01.653431 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.698393 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.698479 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.698502 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.698531 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.698552 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:01Z","lastTransitionTime":"2025-10-07T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.801089 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.801157 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.801166 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.801180 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.801188 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:01Z","lastTransitionTime":"2025-10-07T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.903979 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.904035 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.904053 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.904077 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:01 crc kubenswrapper[4959]: I1007 13:47:01.904093 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:01Z","lastTransitionTime":"2025-10-07T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.007636 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.007691 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.007715 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.007745 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.007765 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:02Z","lastTransitionTime":"2025-10-07T13:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.110669 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.110726 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.110745 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.110768 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.110809 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:02Z","lastTransitionTime":"2025-10-07T13:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.213134 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.213201 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.213223 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.213250 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.213272 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:02Z","lastTransitionTime":"2025-10-07T13:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.315396 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.315456 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.315476 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.315503 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.315520 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:02Z","lastTransitionTime":"2025-10-07T13:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.418402 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.418467 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.418489 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.418517 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.418541 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:02Z","lastTransitionTime":"2025-10-07T13:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.497554 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-metrics-certs\") pod \"network-metrics-daemon-zdps5\" (UID: \"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\") " pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:47:02 crc kubenswrapper[4959]: E1007 13:47:02.497753 4959 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 13:47:02 crc kubenswrapper[4959]: E1007 13:47:02.497915 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-metrics-certs podName:50d4ac5c-0a16-4530-8500-9975eb2b6b5a nodeName:}" failed. No retries permitted until 2025-10-07 13:48:06.497878981 +0000 UTC m=+168.581283346 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-metrics-certs") pod "network-metrics-daemon-zdps5" (UID: "50d4ac5c-0a16-4530-8500-9975eb2b6b5a") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.521277 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.521311 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.521321 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.521336 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.521347 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:02Z","lastTransitionTime":"2025-10-07T13:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.622890 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.622949 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.622971 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.622998 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.623049 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:02Z","lastTransitionTime":"2025-10-07T13:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.653029 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.653088 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:47:02 crc kubenswrapper[4959]: E1007 13:47:02.653245 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.653044 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:47:02 crc kubenswrapper[4959]: E1007 13:47:02.653426 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:47:02 crc kubenswrapper[4959]: E1007 13:47:02.653561 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.726238 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.726284 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.726301 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.726323 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.726339 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:02Z","lastTransitionTime":"2025-10-07T13:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.828714 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.828779 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.828803 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.828832 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.828856 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:02Z","lastTransitionTime":"2025-10-07T13:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.932233 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.932293 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.932312 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.932335 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:02 crc kubenswrapper[4959]: I1007 13:47:02.932351 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:02Z","lastTransitionTime":"2025-10-07T13:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.035403 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.035482 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.035506 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.035539 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.035556 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:03Z","lastTransitionTime":"2025-10-07T13:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.138141 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.138188 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.138200 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.138219 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.138233 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:03Z","lastTransitionTime":"2025-10-07T13:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.240949 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.241015 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.241033 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.241057 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.241075 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:03Z","lastTransitionTime":"2025-10-07T13:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.343071 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.343154 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.343169 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.343189 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.343203 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:03Z","lastTransitionTime":"2025-10-07T13:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.446672 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.446733 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.446749 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.446772 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.446789 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:03Z","lastTransitionTime":"2025-10-07T13:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.549226 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.549289 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.549315 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.549345 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.549368 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:03Z","lastTransitionTime":"2025-10-07T13:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.652595 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.652745 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.652790 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.652807 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.652830 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:03 crc kubenswrapper[4959]: E1007 13:47:03.652839 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.652849 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:03Z","lastTransitionTime":"2025-10-07T13:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.755415 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.755473 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.755490 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.755544 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.755561 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:03Z","lastTransitionTime":"2025-10-07T13:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.858431 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.858493 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.858511 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.858532 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.858549 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:03Z","lastTransitionTime":"2025-10-07T13:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.961192 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.961254 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.961279 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.961309 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:03 crc kubenswrapper[4959]: I1007 13:47:03.961394 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:03Z","lastTransitionTime":"2025-10-07T13:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.064772 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.064822 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.064840 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.064864 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.064882 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:04Z","lastTransitionTime":"2025-10-07T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.168611 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.168679 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.168695 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.168721 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.168740 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:04Z","lastTransitionTime":"2025-10-07T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.271540 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.271606 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.271624 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.271649 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.271668 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:04Z","lastTransitionTime":"2025-10-07T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.374611 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.374671 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.374692 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.374717 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.374735 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:04Z","lastTransitionTime":"2025-10-07T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.477010 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.477077 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.477088 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.477160 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.477175 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:04Z","lastTransitionTime":"2025-10-07T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.580326 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.580416 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.580438 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.580470 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.580492 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:04Z","lastTransitionTime":"2025-10-07T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.653349 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.653427 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.653349 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:47:04 crc kubenswrapper[4959]: E1007 13:47:04.653607 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:47:04 crc kubenswrapper[4959]: E1007 13:47:04.654059 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:47:04 crc kubenswrapper[4959]: E1007 13:47:04.654311 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.655472 4959 scope.go:117] "RemoveContainer" containerID="fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69" Oct 07 13:47:04 crc kubenswrapper[4959]: E1007 13:47:04.656037 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5sfv5_openshift-ovn-kubernetes(b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.683248 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.683332 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.683351 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.683374 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.683393 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:04Z","lastTransitionTime":"2025-10-07T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.787062 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.787160 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.787182 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.787208 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.787226 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:04Z","lastTransitionTime":"2025-10-07T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.890021 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.890058 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.890066 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.890082 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.890092 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:04Z","lastTransitionTime":"2025-10-07T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.993366 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.993434 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.993451 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.993475 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:04 crc kubenswrapper[4959]: I1007 13:47:04.993495 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:04Z","lastTransitionTime":"2025-10-07T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.095888 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.095945 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.095961 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.095985 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.096002 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:05Z","lastTransitionTime":"2025-10-07T13:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.199032 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.199092 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.199203 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.199228 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.199246 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:05Z","lastTransitionTime":"2025-10-07T13:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.301838 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.301886 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.301900 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.301920 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.301931 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:05Z","lastTransitionTime":"2025-10-07T13:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.405096 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.405181 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.405198 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.405226 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.405244 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:05Z","lastTransitionTime":"2025-10-07T13:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.508179 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.508234 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.508254 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.508317 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.508336 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:05Z","lastTransitionTime":"2025-10-07T13:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.611051 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.611129 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.611145 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.611171 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.611191 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:05Z","lastTransitionTime":"2025-10-07T13:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.653178 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:47:05 crc kubenswrapper[4959]: E1007 13:47:05.653348 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.713880 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.713948 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.713972 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.713998 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.714015 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:05Z","lastTransitionTime":"2025-10-07T13:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.816984 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.817040 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.817058 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.817081 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.817097 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:05Z","lastTransitionTime":"2025-10-07T13:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.919619 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.919690 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.919712 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.919869 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:05 crc kubenswrapper[4959]: I1007 13:47:05.919906 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:05Z","lastTransitionTime":"2025-10-07T13:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.022603 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.022637 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.022647 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.022663 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.022674 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:06Z","lastTransitionTime":"2025-10-07T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.126199 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.126255 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.126272 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.126297 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.126314 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:06Z","lastTransitionTime":"2025-10-07T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.229266 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.229333 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.229356 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.229387 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.229412 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:06Z","lastTransitionTime":"2025-10-07T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.333243 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.333304 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.333321 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.333347 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.333364 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:06Z","lastTransitionTime":"2025-10-07T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.432643 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.432735 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.432751 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.432776 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.432793 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:06Z","lastTransitionTime":"2025-10-07T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:06 crc kubenswrapper[4959]: E1007 13:47:06.454750 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:47:06Z is after 2025-08-24T17:21:41Z" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.460222 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.460288 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.460305 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.460749 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.460807 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:06Z","lastTransitionTime":"2025-10-07T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:06 crc kubenswrapper[4959]: E1007 13:47:06.478958 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:47:06Z is after 2025-08-24T17:21:41Z" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.483892 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.483959 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.483982 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.484014 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.484034 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:06Z","lastTransitionTime":"2025-10-07T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:06 crc kubenswrapper[4959]: E1007 13:47:06.502250 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:47:06Z is after 2025-08-24T17:21:41Z" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.506729 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.506802 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.506822 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.506846 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.506866 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:06Z","lastTransitionTime":"2025-10-07T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:06 crc kubenswrapper[4959]: E1007 13:47:06.525812 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:47:06Z is after 2025-08-24T17:21:41Z" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.535660 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.535716 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.535733 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.535757 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.535773 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:06Z","lastTransitionTime":"2025-10-07T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:06 crc kubenswrapper[4959]: E1007 13:47:06.556769 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-07T13:47:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b4282f39-e355-4eb9-8db9-50fa069893da\\\",\\\"systemUUID\\\":\\\"f2d8c8c7-39b4-4371-b614-9cd97280b297\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:47:06Z is after 2025-08-24T17:21:41Z" Oct 07 13:47:06 crc kubenswrapper[4959]: E1007 13:47:06.557010 4959 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.559025 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.559075 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.559155 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.559184 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.559201 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:06Z","lastTransitionTime":"2025-10-07T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.652952 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.653033 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.653033 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:47:06 crc kubenswrapper[4959]: E1007 13:47:06.653169 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:47:06 crc kubenswrapper[4959]: E1007 13:47:06.653328 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:47:06 crc kubenswrapper[4959]: E1007 13:47:06.653472 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.661957 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.662015 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.662035 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.662059 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.662076 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:06Z","lastTransitionTime":"2025-10-07T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.764518 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.764583 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.764609 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.764637 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.764659 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:06Z","lastTransitionTime":"2025-10-07T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.867342 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.867413 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.867437 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.867466 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.867487 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:06Z","lastTransitionTime":"2025-10-07T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.970834 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.970887 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.970906 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.970928 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:06 crc kubenswrapper[4959]: I1007 13:47:06.970944 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:06Z","lastTransitionTime":"2025-10-07T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.074049 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.074088 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.074112 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.074128 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.074139 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:07Z","lastTransitionTime":"2025-10-07T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.176804 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.176879 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.176902 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.176924 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.176941 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:07Z","lastTransitionTime":"2025-10-07T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.280135 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.280171 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.280182 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.280199 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.280211 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:07Z","lastTransitionTime":"2025-10-07T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.382652 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.382699 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.382710 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.382724 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.382733 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:07Z","lastTransitionTime":"2025-10-07T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.485430 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.485477 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.485488 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.485503 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.485512 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:07Z","lastTransitionTime":"2025-10-07T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.587627 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.587673 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.587684 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.587702 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.587714 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:07Z","lastTransitionTime":"2025-10-07T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.652380 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:47:07 crc kubenswrapper[4959]: E1007 13:47:07.652507 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.690461 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.690499 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.690509 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.690523 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.690534 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:07Z","lastTransitionTime":"2025-10-07T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.793072 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.793184 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.793202 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.793226 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.793244 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:07Z","lastTransitionTime":"2025-10-07T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.894970 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.895011 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.895022 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.895036 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.895050 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:07Z","lastTransitionTime":"2025-10-07T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.998179 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.998235 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.998253 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.998274 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:07 crc kubenswrapper[4959]: I1007 13:47:07.998292 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:07Z","lastTransitionTime":"2025-10-07T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.100841 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.100894 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.100905 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.100926 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.100939 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:08Z","lastTransitionTime":"2025-10-07T13:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.203310 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.203350 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.203362 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.203382 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.203394 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:08Z","lastTransitionTime":"2025-10-07T13:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.306450 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.306487 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.306497 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.306515 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.306529 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:08Z","lastTransitionTime":"2025-10-07T13:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.409396 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.409456 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.409482 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.409505 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.409524 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:08Z","lastTransitionTime":"2025-10-07T13:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.513014 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.513081 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.513134 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.513169 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.513192 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:08Z","lastTransitionTime":"2025-10-07T13:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.616414 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.616475 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.616496 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.616521 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.616536 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:08Z","lastTransitionTime":"2025-10-07T13:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.653077 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.653379 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:47:08 crc kubenswrapper[4959]: E1007 13:47:08.653552 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.653573 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:47:08 crc kubenswrapper[4959]: E1007 13:47:08.653744 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:47:08 crc kubenswrapper[4959]: E1007 13:47:08.653964 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.670634 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3c701589735a792ee4928f45cbeba2ac2aaf036bec004394dea02c6d73cb0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lkdmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m5zzg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:47:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.685977 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e2abcf5-d0d4-4ca0-b6a5-db77262adb5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34ab500b1b215518cbfc4602ee0671c6f3f68464467aa392ea5bc38ffb24bd8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1454006ab0b9f83bd85e57a5e7389c81aa81ea5fb4d7c86a6315839de5c37812\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmsrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xpjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:47:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.698800 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zdps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fnzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zdps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:47:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.719591 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.719629 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.719642 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.719660 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.719675 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:08Z","lastTransitionTime":"2025-10-07T13:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.727123 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6641f5dd-1761-4742-8033-2416b6577cf1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecceb9f50fdf3e6652365803d48903bae4d523c2a42f2e3f1bcd7467b76057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://951ef925b63c2a593e48a69b3690c29177519c9d98ef11dd2058e72d58e67398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abea7470986b9c7a6c03042bce1cc2483d0af8df59fa2f5dedffbe58e2f07fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa677a85bcb7ece65cc7d2dc845ef4577394a33d0a5190456b2f95a7cf44ac30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae686eb94ef55f3475bee15ec64748085f134898d8c9387f74a24c072f0fd6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31608ac986e628cf0412fbd8c20ebf8200895d035def7bb8b1e5a2c13e1320eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2696536d15cafb52398e2051be1aa54dd46c13594e85178328308cd04674317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15a057e9d97516effff1c26fe03af3cc58fd5f5cf081fcc5c1ef80715ceb0c3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:47:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.744935 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:47:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.758759 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c521ede959f6b1ec78701d075ca1e52979a8efb851660d847dab6bb07618649\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ba0d500eed75401e1c2589ebbcbf5a848b60bd780d8c13478bcdbe75cb16802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:47:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.772546 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j88pg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0950962-bbba-48a7-8ac8-d6c83e4e61ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9dabbc8b63d02e52dcb7bc11d89170e698364b688a3ef48875575ceb1188c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sgw42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j88pg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:47:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.792181 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4mzl6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2003ec8f-74a0-47bc-8998-0326bfff6e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6898f5d8eb8c5a06853350d8189d1c393cef00392a0c6240b449dbb8617d9f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:46:31Z\\\",\\\"message\\\":\\\"2025-10-07T13:45:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_19b999ee-ab24-4a80-bd1e-5817a041f450\\\\n2025-10-07T13:45:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_19b999ee-ab24-4a80-bd1e-5817a041f450 to /host/opt/cni/bin/\\\\n2025-10-07T13:45:46Z [verbose] multus-daemon started\\\\n2025-10-07T13:45:46Z [verbose] Readiness Indicator file check\\\\n2025-10-07T13:46:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:46:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxqz7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4mzl6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:47:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.814057 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa00eb4-4132-424a-99b9-4e57d30ce059\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c1cbef019f0640b5d51b398d1fc7166c1e253fd33f78f0e50f9754bb1187a6d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acad5e22c53ff5a3c18a7b9c11536b3f41c565356e0cedb32200610e318505a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a108e6423ff51263245ceb5efb815b85b6b3af996580cd9b5840ddc585d50839\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4a8358cc4e14589d938944dabcbf159486576cf6bd58ad4e212765c296d748d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7665bacd797e740b9cf40bb56d3d7ea7cd8312a4d67e96f147e262ff578d3e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7096e4ce671604fc357c134902ec418cf904ee6675cc9726285f7004bb9de5ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6d939e998e8d384688535f0c949e2430c9ec8882eca39895fdb913e1a821e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xfdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vxrtj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:47:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.821233 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.821258 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.821266 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.821278 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.821287 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:08Z","lastTransitionTime":"2025-10-07T13:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.827087 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-km89w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c27a0cb-3867-4c31-b438-c34570f7e1cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e00af68ccf268d6752f8423f12467a1be85585e8f8a04ccb1e66328170c411f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s45mf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-km89w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:47:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.839028 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1ac79b-8fe3-40cd-925d-d840f0995abf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae966401c1e804da8649ca9c014cfc2eed8dbb12a0576cf90b6c7a3526513601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6d82be64b945bd47d6149c695f43152998644d31f0103bfff861199d5efb191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6d82be64b945bd47d6149c695f43152998644d31f0103bfff861199d5efb191\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:47:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.855008 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3ea9e08-bdcd-4b83-bffb-4a0fff064de6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d7236a95c0e0afc629b2381cff61f30367768b0c91cd01dff919dc6041f8544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7260e2d07cbcf5a939f528d9fb71a5755622d85f993994da46192927962b1a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528275c15a33367e98901a725b4db3bcdb2aba40a59fc7d7bfbfb95d39cfdc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f887e869b1507ecede53bb7667e70b29a730e55e74171ee16ff5c0a2247ed80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cec56ce1b9d2007df1929849276f95e573a91269d30180d467b5aeb11a3bff58\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-07T13:45:32Z\\\",\\\"message\\\":\\\"W1007 13:45:21.908461 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1007 13:45:21.908719 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759844721 cert, and key in /tmp/serving-cert-1247860539/serving-signer.crt, /tmp/serving-cert-1247860539/serving-signer.key\\\\nI1007 13:45:22.175274 1 observer_polling.go:159] Starting file observer\\\\nW1007 13:45:22.177643 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1007 13:45:22.177898 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1007 13:45:22.182216 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1247860539/tls.crt::/tmp/serving-cert-1247860539/tls.key\\\\\\\"\\\\nF1007 13:45:32.489362 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea42ce0064363af9c365ebe894d1cfc14c217ea095659d0f7ead7dc6ebfccef0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d02b1c9aea5db329591150b0d80b6f322bd8024e30c7d9d3c08f2d5cfbc9669c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:47:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.868218 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2133e3aa837c13035a01dc5b5b83f70318dc4a2cc8dc5947582ca64982e5a061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:47:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.884493 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:47:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.909828 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-07T13:46:36Z\\\",\\\"message\\\":\\\": current time 2025-10-07T13:46:36Z is after 2025-08-24T17:21:41Z]\\\\nI1007 13:46:36.558255 6945 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-diagnostics/network-check-target]} name:Service_openshift-network-diagnostics/network-check-target_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.219:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7594bb65-e742-44b3-a975-d639b1128be5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1007 13:46:36.558260 6945 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-apiserver-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"8b82f026-5975-4a1b-bb18-08d5d51147ec\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-07T13:46:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5sfv5_openshift-ovn-kubernetes(b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-969lv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5sfv5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:47:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.923501 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.923536 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.923547 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.923564 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.923577 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:08Z","lastTransitionTime":"2025-10-07T13:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.927167 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e325099f-e0ff-433d-bef7-6909c5c86bea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3ecb7da009b7ed9e5c0412125be08cf486bf4f0e74480f2843877b0816944fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6dab7bc77a409dca97cffd302d0086021bf51146e8ec69a53cbddb87a4026aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b500df391c3b1ef8f9689a04e58c19e603667507a68fb8d65ea768690425085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://957f312e600987fd82f8ba93fd4aebc2d696d76ff8486a0bd93cb34ca21f7b3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-07T13:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:47:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.945142 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"963f5c32-ea27-42c3-8f4a-5b7100e2299f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e648b1ce83f27791a550d7a40678083ec5eab952c5c0813ed30ef32b148a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c5f206359d97344ad314d9a9eac0882df1c4b48aa0d771e49c7fe70bd5d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3901e035ed3c9be2d5df3db927678c1a29c8aa07ef880f0c11b101f33a6fd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7734d46aaafacaa98f26499c03fde0505bc58bb4d1256829b908a2842571e7a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-07T13:45:18Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:47:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.963829 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6a38de8aa1d4536a73524ef085f5ffadee0c244b8b934d5bd9b6e27ce49ba78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-07T13:45:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:47:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:47:08 crc kubenswrapper[4959]: I1007 13:47:08.984997 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-07T13:45:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-07T13:47:08Z is after 2025-08-24T17:21:41Z" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.026088 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.026437 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.026587 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.026745 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.026885 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:09Z","lastTransitionTime":"2025-10-07T13:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.130126 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.130398 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.130484 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.130559 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.130638 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:09Z","lastTransitionTime":"2025-10-07T13:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.232624 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.232856 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.232945 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.233037 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.233132 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:09Z","lastTransitionTime":"2025-10-07T13:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.335162 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.335228 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.335247 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.335270 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.335287 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:09Z","lastTransitionTime":"2025-10-07T13:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.438096 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.438376 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.438446 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.438504 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.438558 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:09Z","lastTransitionTime":"2025-10-07T13:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.540801 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.540838 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.540848 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.540863 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.540874 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:09Z","lastTransitionTime":"2025-10-07T13:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.643017 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.643140 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.643176 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.643206 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.643229 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:09Z","lastTransitionTime":"2025-10-07T13:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.653081 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:47:09 crc kubenswrapper[4959]: E1007 13:47:09.653352 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.746389 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.746431 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.746445 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.746464 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.746474 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:09Z","lastTransitionTime":"2025-10-07T13:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.849614 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.849664 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.849674 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.849690 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.849701 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:09Z","lastTransitionTime":"2025-10-07T13:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.952833 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.952902 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.952919 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.952944 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:09 crc kubenswrapper[4959]: I1007 13:47:09.952961 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:09Z","lastTransitionTime":"2025-10-07T13:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.055880 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.055923 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.055932 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.055946 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.055955 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:10Z","lastTransitionTime":"2025-10-07T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.159263 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.159342 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.159363 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.159385 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.159405 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:10Z","lastTransitionTime":"2025-10-07T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.261847 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.261890 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.261899 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.261913 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.261926 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:10Z","lastTransitionTime":"2025-10-07T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.365327 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.365433 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.365454 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.365514 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.365533 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:10Z","lastTransitionTime":"2025-10-07T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.468419 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.468508 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.468531 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.468881 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.468903 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:10Z","lastTransitionTime":"2025-10-07T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.571555 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.571598 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.571610 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.571626 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.571639 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:10Z","lastTransitionTime":"2025-10-07T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.653383 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.653508 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.653537 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:47:10 crc kubenswrapper[4959]: E1007 13:47:10.653639 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:47:10 crc kubenswrapper[4959]: E1007 13:47:10.653805 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:47:10 crc kubenswrapper[4959]: E1007 13:47:10.653901 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.674196 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.674256 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.674276 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.674307 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.674330 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:10Z","lastTransitionTime":"2025-10-07T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.777663 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.777737 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.777757 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.777782 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.777799 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:10Z","lastTransitionTime":"2025-10-07T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.880599 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.880664 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.880683 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.880707 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.880728 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:10Z","lastTransitionTime":"2025-10-07T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.983880 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.983918 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.983927 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.983941 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:10 crc kubenswrapper[4959]: I1007 13:47:10.983952 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:10Z","lastTransitionTime":"2025-10-07T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.087470 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.087548 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.087569 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.087593 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.087611 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:11Z","lastTransitionTime":"2025-10-07T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.190655 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.190711 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.190728 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.190753 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.190771 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:11Z","lastTransitionTime":"2025-10-07T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.294080 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.294215 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.294241 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.294272 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.294290 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:11Z","lastTransitionTime":"2025-10-07T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.397869 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.397917 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.397986 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.398014 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.398067 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:11Z","lastTransitionTime":"2025-10-07T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.501200 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.501258 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.501276 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.501304 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.501321 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:11Z","lastTransitionTime":"2025-10-07T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.603915 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.603967 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.603981 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.604002 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.604016 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:11Z","lastTransitionTime":"2025-10-07T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.652944 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:47:11 crc kubenswrapper[4959]: E1007 13:47:11.653213 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.707025 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.707080 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.707157 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.707182 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.707201 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:11Z","lastTransitionTime":"2025-10-07T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.811342 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.811447 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.811505 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.811531 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.811603 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:11Z","lastTransitionTime":"2025-10-07T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.915185 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.915272 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.915295 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.915325 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:11 crc kubenswrapper[4959]: I1007 13:47:11.915344 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:11Z","lastTransitionTime":"2025-10-07T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.018214 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.018291 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.018309 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.018336 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.018353 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:12Z","lastTransitionTime":"2025-10-07T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.121449 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.121522 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.121546 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.121584 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.121609 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:12Z","lastTransitionTime":"2025-10-07T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.225010 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.225068 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.225087 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.225153 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.225175 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:12Z","lastTransitionTime":"2025-10-07T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.328581 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.328644 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.328669 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.328700 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.328722 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:12Z","lastTransitionTime":"2025-10-07T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.438150 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.438257 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.438289 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.438333 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.438369 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:12Z","lastTransitionTime":"2025-10-07T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.542085 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.542189 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.542209 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.542235 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.542254 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:12Z","lastTransitionTime":"2025-10-07T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.646619 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.646682 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.646698 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.646724 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.646742 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:12Z","lastTransitionTime":"2025-10-07T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.653654 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.653675 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.653845 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:47:12 crc kubenswrapper[4959]: E1007 13:47:12.654065 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:47:12 crc kubenswrapper[4959]: E1007 13:47:12.654258 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:47:12 crc kubenswrapper[4959]: E1007 13:47:12.654407 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.750929 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.750998 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.751019 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.751051 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.751073 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:12Z","lastTransitionTime":"2025-10-07T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.854879 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.854948 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.854968 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.854999 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.855022 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:12Z","lastTransitionTime":"2025-10-07T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.959156 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.959223 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.959245 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.959274 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:12 crc kubenswrapper[4959]: I1007 13:47:12.959293 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:12Z","lastTransitionTime":"2025-10-07T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.063011 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.063061 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.063079 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.063139 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.063158 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:13Z","lastTransitionTime":"2025-10-07T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.166597 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.166671 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.166691 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.166720 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.166738 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:13Z","lastTransitionTime":"2025-10-07T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.269481 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.269561 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.269581 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.269611 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.269632 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:13Z","lastTransitionTime":"2025-10-07T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.373029 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.373164 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.373190 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.373224 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.373251 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:13Z","lastTransitionTime":"2025-10-07T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.476599 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.476666 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.476688 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.476717 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.476738 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:13Z","lastTransitionTime":"2025-10-07T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.608068 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.608172 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.608197 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.608226 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.608246 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:13Z","lastTransitionTime":"2025-10-07T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.653411 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:47:13 crc kubenswrapper[4959]: E1007 13:47:13.653866 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.711607 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.711674 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.711691 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.711716 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.711737 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:13Z","lastTransitionTime":"2025-10-07T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.815197 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.815280 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.815307 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.815333 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.815356 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:13Z","lastTransitionTime":"2025-10-07T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.918551 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.918607 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.918624 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.918650 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:13 crc kubenswrapper[4959]: I1007 13:47:13.918668 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:13Z","lastTransitionTime":"2025-10-07T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.022228 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.022287 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.022304 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.022350 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.022368 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:14Z","lastTransitionTime":"2025-10-07T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.125513 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.125569 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.125586 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.125610 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.125635 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:14Z","lastTransitionTime":"2025-10-07T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.228341 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.228409 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.228426 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.228455 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.228473 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:14Z","lastTransitionTime":"2025-10-07T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.332746 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.332839 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.332862 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.332895 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.332919 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:14Z","lastTransitionTime":"2025-10-07T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.436692 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.436769 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.436786 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.436812 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.436832 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:14Z","lastTransitionTime":"2025-10-07T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.540279 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.540378 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.540403 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.540468 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.540491 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:14Z","lastTransitionTime":"2025-10-07T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.643761 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.643830 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.643848 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.643873 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.643892 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:14Z","lastTransitionTime":"2025-10-07T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.654296 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.654421 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:47:14 crc kubenswrapper[4959]: E1007 13:47:14.654496 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.654692 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:47:14 crc kubenswrapper[4959]: E1007 13:47:14.654915 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:47:14 crc kubenswrapper[4959]: E1007 13:47:14.655058 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.748159 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.748764 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.748783 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.748812 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.748830 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:14Z","lastTransitionTime":"2025-10-07T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.851940 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.851993 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.852013 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.852038 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.852057 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:14Z","lastTransitionTime":"2025-10-07T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.956058 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.956197 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.956218 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.956247 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:14 crc kubenswrapper[4959]: I1007 13:47:14.956269 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:14Z","lastTransitionTime":"2025-10-07T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.060287 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.060745 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.060892 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.061034 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.061238 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:15Z","lastTransitionTime":"2025-10-07T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.164568 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.164939 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.165078 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.165279 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.165421 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:15Z","lastTransitionTime":"2025-10-07T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.268544 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.268628 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.268648 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.268676 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.268694 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:15Z","lastTransitionTime":"2025-10-07T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.374192 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.374254 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.374274 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.374299 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.374316 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:15Z","lastTransitionTime":"2025-10-07T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.477846 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.477916 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.477935 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.477961 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.477980 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:15Z","lastTransitionTime":"2025-10-07T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.581458 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.581524 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.581542 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.581565 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.581585 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:15Z","lastTransitionTime":"2025-10-07T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.653060 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:47:15 crc kubenswrapper[4959]: E1007 13:47:15.653304 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.684698 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.684749 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.684767 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.684790 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.684807 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:15Z","lastTransitionTime":"2025-10-07T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.788611 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.788685 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.788703 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.788728 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.788749 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:15Z","lastTransitionTime":"2025-10-07T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.891932 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.891999 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.892022 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.892059 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.892084 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:15Z","lastTransitionTime":"2025-10-07T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.995154 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.995207 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.995219 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.995236 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:15 crc kubenswrapper[4959]: I1007 13:47:15.995248 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:15Z","lastTransitionTime":"2025-10-07T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.098644 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.098680 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.098694 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.098712 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.098724 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:16Z","lastTransitionTime":"2025-10-07T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.209758 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.209808 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.209820 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.209836 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.209849 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:16Z","lastTransitionTime":"2025-10-07T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.313480 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.313539 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.313558 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.313582 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.313600 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:16Z","lastTransitionTime":"2025-10-07T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.417637 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.417710 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.417729 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.417757 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.417778 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:16Z","lastTransitionTime":"2025-10-07T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.521073 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.521189 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.521219 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.521253 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.521281 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:16Z","lastTransitionTime":"2025-10-07T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.623837 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.623908 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.623930 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.623961 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.623982 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:16Z","lastTransitionTime":"2025-10-07T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.653604 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.653727 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.653798 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:47:16 crc kubenswrapper[4959]: E1007 13:47:16.653803 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:47:16 crc kubenswrapper[4959]: E1007 13:47:16.653899 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:47:16 crc kubenswrapper[4959]: E1007 13:47:16.654053 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.727316 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.727384 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.727401 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.727427 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.727445 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:16Z","lastTransitionTime":"2025-10-07T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.759468 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.759532 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.759548 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.759568 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.759585 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-07T13:47:16Z","lastTransitionTime":"2025-10-07T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.846645 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-gmq7h"] Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.847430 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gmq7h" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.850503 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.853024 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.853565 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.853818 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.906414 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-4mzl6" podStartSLOduration=93.90637177 podStartE2EDuration="1m33.90637177s" podCreationTimestamp="2025-10-07 13:45:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:16.876055341 +0000 UTC m=+118.959459706" watchObservedRunningTime="2025-10-07 13:47:16.90637177 +0000 UTC m=+118.989776135" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.931704 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-vxrtj" podStartSLOduration=93.931664341 podStartE2EDuration="1m33.931664341s" podCreationTimestamp="2025-10-07 13:45:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:16.912552926 +0000 UTC m=+118.995957251" watchObservedRunningTime="2025-10-07 13:47:16.931664341 +0000 UTC m=+119.015068706" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.932043 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-km89w" podStartSLOduration=93.932033171 podStartE2EDuration="1m33.932033171s" podCreationTimestamp="2025-10-07 13:45:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:16.93161653 +0000 UTC m=+119.015020895" watchObservedRunningTime="2025-10-07 13:47:16.932033171 +0000 UTC m=+119.015437536" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.946618 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=45.946582910000004 podStartE2EDuration="45.94658291s" podCreationTimestamp="2025-10-07 13:46:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:16.945989735 +0000 UTC m=+119.029394090" watchObservedRunningTime="2025-10-07 13:47:16.94658291 +0000 UTC m=+119.029987265" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.961557 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d6cf5489-0607-4e1b-b6a6-b2110f5a5a38-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-gmq7h\" (UID: \"d6cf5489-0607-4e1b-b6a6-b2110f5a5a38\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gmq7h" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.962723 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/d6cf5489-0607-4e1b-b6a6-b2110f5a5a38-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-gmq7h\" (UID: \"d6cf5489-0607-4e1b-b6a6-b2110f5a5a38\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gmq7h" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.962935 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6cf5489-0607-4e1b-b6a6-b2110f5a5a38-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-gmq7h\" (UID: \"d6cf5489-0607-4e1b-b6a6-b2110f5a5a38\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gmq7h" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.963178 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/d6cf5489-0607-4e1b-b6a6-b2110f5a5a38-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-gmq7h\" (UID: \"d6cf5489-0607-4e1b-b6a6-b2110f5a5a38\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gmq7h" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.963573 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d6cf5489-0607-4e1b-b6a6-b2110f5a5a38-service-ca\") pod \"cluster-version-operator-5c965bbfc6-gmq7h\" (UID: \"d6cf5489-0607-4e1b-b6a6-b2110f5a5a38\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gmq7h" Oct 07 13:47:16 crc kubenswrapper[4959]: I1007 13:47:16.993537 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=98.99350013 podStartE2EDuration="1m38.99350013s" podCreationTimestamp="2025-10-07 13:45:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:16.980065559 +0000 UTC m=+119.063469924" watchObservedRunningTime="2025-10-07 13:47:16.99350013 +0000 UTC m=+119.076904495" Oct 07 13:47:17 crc kubenswrapper[4959]: I1007 13:47:17.064906 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d6cf5489-0607-4e1b-b6a6-b2110f5a5a38-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-gmq7h\" (UID: \"d6cf5489-0607-4e1b-b6a6-b2110f5a5a38\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gmq7h" Oct 07 13:47:17 crc kubenswrapper[4959]: I1007 13:47:17.064979 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/d6cf5489-0607-4e1b-b6a6-b2110f5a5a38-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-gmq7h\" (UID: \"d6cf5489-0607-4e1b-b6a6-b2110f5a5a38\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gmq7h" Oct 07 13:47:17 crc kubenswrapper[4959]: I1007 13:47:17.065016 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6cf5489-0607-4e1b-b6a6-b2110f5a5a38-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-gmq7h\" (UID: \"d6cf5489-0607-4e1b-b6a6-b2110f5a5a38\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gmq7h" Oct 07 13:47:17 crc kubenswrapper[4959]: I1007 13:47:17.065065 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/d6cf5489-0607-4e1b-b6a6-b2110f5a5a38-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-gmq7h\" (UID: \"d6cf5489-0607-4e1b-b6a6-b2110f5a5a38\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gmq7h" Oct 07 13:47:17 crc kubenswrapper[4959]: I1007 13:47:17.065139 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d6cf5489-0607-4e1b-b6a6-b2110f5a5a38-service-ca\") pod \"cluster-version-operator-5c965bbfc6-gmq7h\" (UID: \"d6cf5489-0607-4e1b-b6a6-b2110f5a5a38\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gmq7h" Oct 07 13:47:17 crc kubenswrapper[4959]: I1007 13:47:17.065687 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/d6cf5489-0607-4e1b-b6a6-b2110f5a5a38-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-gmq7h\" (UID: \"d6cf5489-0607-4e1b-b6a6-b2110f5a5a38\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gmq7h" Oct 07 13:47:17 crc kubenswrapper[4959]: I1007 13:47:17.066238 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/d6cf5489-0607-4e1b-b6a6-b2110f5a5a38-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-gmq7h\" (UID: \"d6cf5489-0607-4e1b-b6a6-b2110f5a5a38\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gmq7h" Oct 07 13:47:17 crc kubenswrapper[4959]: I1007 13:47:17.067066 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d6cf5489-0607-4e1b-b6a6-b2110f5a5a38-service-ca\") pod \"cluster-version-operator-5c965bbfc6-gmq7h\" (UID: \"d6cf5489-0607-4e1b-b6a6-b2110f5a5a38\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gmq7h" Oct 07 13:47:17 crc kubenswrapper[4959]: I1007 13:47:17.078366 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6cf5489-0607-4e1b-b6a6-b2110f5a5a38-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-gmq7h\" (UID: \"d6cf5489-0607-4e1b-b6a6-b2110f5a5a38\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gmq7h" Oct 07 13:47:17 crc kubenswrapper[4959]: I1007 13:47:17.093924 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d6cf5489-0607-4e1b-b6a6-b2110f5a5a38-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-gmq7h\" (UID: \"d6cf5489-0607-4e1b-b6a6-b2110f5a5a38\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gmq7h" Oct 07 13:47:17 crc kubenswrapper[4959]: I1007 13:47:17.106390 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=69.106366272 podStartE2EDuration="1m9.106366272s" podCreationTimestamp="2025-10-07 13:46:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:17.105676765 +0000 UTC m=+119.189081100" watchObservedRunningTime="2025-10-07 13:47:17.106366272 +0000 UTC m=+119.189770607" Oct 07 13:47:17 crc kubenswrapper[4959]: I1007 13:47:17.125328 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=92.125303993 podStartE2EDuration="1m32.125303993s" podCreationTimestamp="2025-10-07 13:45:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:17.124877712 +0000 UTC m=+119.208282027" watchObservedRunningTime="2025-10-07 13:47:17.125303993 +0000 UTC m=+119.208708328" Oct 07 13:47:17 crc kubenswrapper[4959]: I1007 13:47:17.162219 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podStartSLOduration=94.162198038 podStartE2EDuration="1m34.162198038s" podCreationTimestamp="2025-10-07 13:45:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:17.161793998 +0000 UTC m=+119.245198333" watchObservedRunningTime="2025-10-07 13:47:17.162198038 +0000 UTC m=+119.245602373" Oct 07 13:47:17 crc kubenswrapper[4959]: I1007 13:47:17.171978 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gmq7h" Oct 07 13:47:17 crc kubenswrapper[4959]: I1007 13:47:17.197889 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xpjs" podStartSLOduration=93.197869043 podStartE2EDuration="1m33.197869043s" podCreationTimestamp="2025-10-07 13:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:17.179137468 +0000 UTC m=+119.262541803" watchObservedRunningTime="2025-10-07 13:47:17.197869043 +0000 UTC m=+119.281273368" Oct 07 13:47:17 crc kubenswrapper[4959]: I1007 13:47:17.226268 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=95.226243613 podStartE2EDuration="1m35.226243613s" podCreationTimestamp="2025-10-07 13:45:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:17.224030727 +0000 UTC m=+119.307435062" watchObservedRunningTime="2025-10-07 13:47:17.226243613 +0000 UTC m=+119.309647948" Oct 07 13:47:17 crc kubenswrapper[4959]: I1007 13:47:17.265001 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-j88pg" podStartSLOduration=94.264980585 podStartE2EDuration="1m34.264980585s" podCreationTimestamp="2025-10-07 13:45:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:17.264795971 +0000 UTC m=+119.348200296" watchObservedRunningTime="2025-10-07 13:47:17.264980585 +0000 UTC m=+119.348384910" Oct 07 13:47:17 crc kubenswrapper[4959]: I1007 13:47:17.273698 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gmq7h" event={"ID":"d6cf5489-0607-4e1b-b6a6-b2110f5a5a38","Type":"ContainerStarted","Data":"3c6d7d366ab903b6407bc54a08a7cb7a436f0ca59e84881631223f0c32acd3c1"} Oct 07 13:47:17 crc kubenswrapper[4959]: I1007 13:47:17.652996 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:47:17 crc kubenswrapper[4959]: E1007 13:47:17.654079 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:47:17 crc kubenswrapper[4959]: I1007 13:47:17.654236 4959 scope.go:117] "RemoveContainer" containerID="fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69" Oct 07 13:47:18 crc kubenswrapper[4959]: I1007 13:47:18.283503 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4mzl6_2003ec8f-74a0-47bc-8998-0326bfff6e7f/kube-multus/1.log" Oct 07 13:47:18 crc kubenswrapper[4959]: I1007 13:47:18.284263 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4mzl6_2003ec8f-74a0-47bc-8998-0326bfff6e7f/kube-multus/0.log" Oct 07 13:47:18 crc kubenswrapper[4959]: I1007 13:47:18.284338 4959 generic.go:334] "Generic (PLEG): container finished" podID="2003ec8f-74a0-47bc-8998-0326bfff6e7f" containerID="6898f5d8eb8c5a06853350d8189d1c393cef00392a0c6240b449dbb8617d9f2f" exitCode=1 Oct 07 13:47:18 crc kubenswrapper[4959]: I1007 13:47:18.284432 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4mzl6" event={"ID":"2003ec8f-74a0-47bc-8998-0326bfff6e7f","Type":"ContainerDied","Data":"6898f5d8eb8c5a06853350d8189d1c393cef00392a0c6240b449dbb8617d9f2f"} Oct 07 13:47:18 crc kubenswrapper[4959]: I1007 13:47:18.284489 4959 scope.go:117] "RemoveContainer" containerID="9da9a1b0a755e835de1d4482ff6ce2e7cf8db092492bd6aa567a2877c2897ebb" Oct 07 13:47:18 crc kubenswrapper[4959]: I1007 13:47:18.285043 4959 scope.go:117] "RemoveContainer" containerID="6898f5d8eb8c5a06853350d8189d1c393cef00392a0c6240b449dbb8617d9f2f" Oct 07 13:47:18 crc kubenswrapper[4959]: E1007 13:47:18.285266 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-4mzl6_openshift-multus(2003ec8f-74a0-47bc-8998-0326bfff6e7f)\"" pod="openshift-multus/multus-4mzl6" podUID="2003ec8f-74a0-47bc-8998-0326bfff6e7f" Oct 07 13:47:18 crc kubenswrapper[4959]: I1007 13:47:18.289521 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gmq7h" event={"ID":"d6cf5489-0607-4e1b-b6a6-b2110f5a5a38","Type":"ContainerStarted","Data":"d3c24dcea1587f0163a6bf88f09ea670c03377f61233beb1aea8eff6816f8b22"} Oct 07 13:47:18 crc kubenswrapper[4959]: I1007 13:47:18.295270 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5sfv5_b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b/ovnkube-controller/3.log" Oct 07 13:47:18 crc kubenswrapper[4959]: I1007 13:47:18.299075 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" event={"ID":"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b","Type":"ContainerStarted","Data":"d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825"} Oct 07 13:47:18 crc kubenswrapper[4959]: I1007 13:47:18.300216 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:47:18 crc kubenswrapper[4959]: I1007 13:47:18.363460 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" podStartSLOduration=94.363445357 podStartE2EDuration="1m34.363445357s" podCreationTimestamp="2025-10-07 13:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:18.362220936 +0000 UTC m=+120.445625261" watchObservedRunningTime="2025-10-07 13:47:18.363445357 +0000 UTC m=+120.446849702" Oct 07 13:47:18 crc kubenswrapper[4959]: I1007 13:47:18.653412 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:47:18 crc kubenswrapper[4959]: I1007 13:47:18.653618 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:47:18 crc kubenswrapper[4959]: E1007 13:47:18.654457 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:47:18 crc kubenswrapper[4959]: I1007 13:47:18.654542 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:47:18 crc kubenswrapper[4959]: E1007 13:47:18.654735 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:47:18 crc kubenswrapper[4959]: E1007 13:47:18.654955 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:47:18 crc kubenswrapper[4959]: E1007 13:47:18.686272 4959 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 07 13:47:18 crc kubenswrapper[4959]: I1007 13:47:18.729880 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gmq7h" podStartSLOduration=95.7298476 podStartE2EDuration="1m35.7298476s" podCreationTimestamp="2025-10-07 13:45:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:18.376328553 +0000 UTC m=+120.459732878" watchObservedRunningTime="2025-10-07 13:47:18.7298476 +0000 UTC m=+120.813251965" Oct 07 13:47:18 crc kubenswrapper[4959]: I1007 13:47:18.730916 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-zdps5"] Oct 07 13:47:18 crc kubenswrapper[4959]: E1007 13:47:18.757279 4959 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 07 13:47:19 crc kubenswrapper[4959]: I1007 13:47:19.305050 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4mzl6_2003ec8f-74a0-47bc-8998-0326bfff6e7f/kube-multus/1.log" Oct 07 13:47:19 crc kubenswrapper[4959]: I1007 13:47:19.305223 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:47:19 crc kubenswrapper[4959]: E1007 13:47:19.305560 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:47:19 crc kubenswrapper[4959]: I1007 13:47:19.652753 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:47:19 crc kubenswrapper[4959]: E1007 13:47:19.652961 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:47:20 crc kubenswrapper[4959]: I1007 13:47:20.653301 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:47:20 crc kubenswrapper[4959]: I1007 13:47:20.653387 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:47:20 crc kubenswrapper[4959]: E1007 13:47:20.653455 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:47:20 crc kubenswrapper[4959]: E1007 13:47:20.653558 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:47:21 crc kubenswrapper[4959]: I1007 13:47:21.653433 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:47:21 crc kubenswrapper[4959]: I1007 13:47:21.653452 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:47:21 crc kubenswrapper[4959]: E1007 13:47:21.653643 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:47:21 crc kubenswrapper[4959]: E1007 13:47:21.653852 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:47:22 crc kubenswrapper[4959]: I1007 13:47:22.652924 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:47:22 crc kubenswrapper[4959]: E1007 13:47:22.653132 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:47:22 crc kubenswrapper[4959]: I1007 13:47:22.653336 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:47:22 crc kubenswrapper[4959]: E1007 13:47:22.653688 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:47:23 crc kubenswrapper[4959]: I1007 13:47:23.652643 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:47:23 crc kubenswrapper[4959]: I1007 13:47:23.652670 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:47:23 crc kubenswrapper[4959]: E1007 13:47:23.652874 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:47:23 crc kubenswrapper[4959]: E1007 13:47:23.653129 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:47:23 crc kubenswrapper[4959]: E1007 13:47:23.758751 4959 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 07 13:47:24 crc kubenswrapper[4959]: I1007 13:47:24.653502 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:47:24 crc kubenswrapper[4959]: I1007 13:47:24.653521 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:47:24 crc kubenswrapper[4959]: E1007 13:47:24.653689 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:47:24 crc kubenswrapper[4959]: E1007 13:47:24.653831 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:47:25 crc kubenswrapper[4959]: I1007 13:47:25.653378 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:47:25 crc kubenswrapper[4959]: I1007 13:47:25.653464 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:47:25 crc kubenswrapper[4959]: E1007 13:47:25.653634 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:47:25 crc kubenswrapper[4959]: E1007 13:47:25.653776 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:47:26 crc kubenswrapper[4959]: I1007 13:47:26.652613 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:47:26 crc kubenswrapper[4959]: I1007 13:47:26.652720 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:47:26 crc kubenswrapper[4959]: E1007 13:47:26.653195 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:47:26 crc kubenswrapper[4959]: E1007 13:47:26.653365 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:47:27 crc kubenswrapper[4959]: I1007 13:47:27.653077 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:47:27 crc kubenswrapper[4959]: I1007 13:47:27.653121 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:47:27 crc kubenswrapper[4959]: E1007 13:47:27.653293 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:47:27 crc kubenswrapper[4959]: E1007 13:47:27.653466 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:47:28 crc kubenswrapper[4959]: I1007 13:47:28.653294 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:47:28 crc kubenswrapper[4959]: I1007 13:47:28.653367 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:47:28 crc kubenswrapper[4959]: E1007 13:47:28.655082 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:47:28 crc kubenswrapper[4959]: E1007 13:47:28.655353 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:47:28 crc kubenswrapper[4959]: E1007 13:47:28.759860 4959 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 07 13:47:29 crc kubenswrapper[4959]: I1007 13:47:29.653037 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:47:29 crc kubenswrapper[4959]: I1007 13:47:29.653057 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:47:29 crc kubenswrapper[4959]: E1007 13:47:29.653332 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:47:29 crc kubenswrapper[4959]: E1007 13:47:29.653844 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:47:30 crc kubenswrapper[4959]: I1007 13:47:30.652924 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:47:30 crc kubenswrapper[4959]: I1007 13:47:30.652939 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:47:30 crc kubenswrapper[4959]: E1007 13:47:30.653429 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:47:30 crc kubenswrapper[4959]: E1007 13:47:30.653522 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:47:30 crc kubenswrapper[4959]: I1007 13:47:30.653664 4959 scope.go:117] "RemoveContainer" containerID="6898f5d8eb8c5a06853350d8189d1c393cef00392a0c6240b449dbb8617d9f2f" Oct 07 13:47:31 crc kubenswrapper[4959]: I1007 13:47:31.355074 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4mzl6_2003ec8f-74a0-47bc-8998-0326bfff6e7f/kube-multus/1.log" Oct 07 13:47:31 crc kubenswrapper[4959]: I1007 13:47:31.355235 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4mzl6" event={"ID":"2003ec8f-74a0-47bc-8998-0326bfff6e7f","Type":"ContainerStarted","Data":"c030ef3529fc036b6e27c94b7c4c51d59bef679992157f58e1acf0858c581ed1"} Oct 07 13:47:31 crc kubenswrapper[4959]: I1007 13:47:31.652763 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:47:31 crc kubenswrapper[4959]: E1007 13:47:31.653028 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:47:31 crc kubenswrapper[4959]: I1007 13:47:31.652772 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:47:31 crc kubenswrapper[4959]: E1007 13:47:31.653208 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:47:32 crc kubenswrapper[4959]: I1007 13:47:32.653616 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:47:32 crc kubenswrapper[4959]: I1007 13:47:32.653821 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:47:32 crc kubenswrapper[4959]: E1007 13:47:32.654021 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 07 13:47:32 crc kubenswrapper[4959]: E1007 13:47:32.654317 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 07 13:47:33 crc kubenswrapper[4959]: I1007 13:47:33.653410 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:47:33 crc kubenswrapper[4959]: I1007 13:47:33.653451 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:47:33 crc kubenswrapper[4959]: E1007 13:47:33.653567 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zdps5" podUID="50d4ac5c-0a16-4530-8500-9975eb2b6b5a" Oct 07 13:47:33 crc kubenswrapper[4959]: E1007 13:47:33.653639 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 07 13:47:34 crc kubenswrapper[4959]: I1007 13:47:34.652876 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:47:34 crc kubenswrapper[4959]: I1007 13:47:34.652935 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:47:34 crc kubenswrapper[4959]: I1007 13:47:34.656223 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 07 13:47:34 crc kubenswrapper[4959]: I1007 13:47:34.656258 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 07 13:47:34 crc kubenswrapper[4959]: I1007 13:47:34.658987 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 07 13:47:34 crc kubenswrapper[4959]: I1007 13:47:34.659211 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 07 13:47:35 crc kubenswrapper[4959]: I1007 13:47:35.653258 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:47:35 crc kubenswrapper[4959]: I1007 13:47:35.653304 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:47:35 crc kubenswrapper[4959]: I1007 13:47:35.656294 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 07 13:47:35 crc kubenswrapper[4959]: I1007 13:47:35.657540 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.288851 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.346676 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.347395 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.347857 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9drmk"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.348278 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9drmk" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.350175 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.352323 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.353209 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.354472 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.355702 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.360635 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.360642 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.360732 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.361151 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.361848 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.363803 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.364067 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.364171 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.364525 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.364927 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.364971 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-67jl7"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.365506 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.367915 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-l78hd"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.368448 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l78hd" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.368706 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.368967 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.369578 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.369645 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.369777 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.369926 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.374131 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.374271 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.374546 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.374582 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.375066 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.375255 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.375327 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.375540 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.377214 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-82gdf"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.377738 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-82gdf" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.380044 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-cw2cl"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.381331 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-cw2cl" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.381983 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zg2sm"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.382640 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.382652 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zg2sm" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.382778 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.383352 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.383545 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.383742 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.383910 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.383998 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.384252 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.384528 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.384833 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.384956 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.386753 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qr2pg"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.387289 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-qr2pg" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.387536 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-tv7sf"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.393288 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.395406 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.395923 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.397664 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-dt2k4"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.398178 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8k48v"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.398389 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-phxkp"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.398609 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mx6p7"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.414470 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mx6p7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.414627 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpw6q\" (UniqueName: \"kubernetes.io/projected/ced32e5c-2931-4a96-a779-1c1c6b71243e-kube-api-access-tpw6q\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.414866 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbwpg\" (UniqueName: \"kubernetes.io/projected/cc0f9f33-98c7-433b-a2e8-2a1dd0066c97-kube-api-access-nbwpg\") pod \"apiserver-7bbb656c7d-hsdts\" (UID: \"cc0f9f33-98c7-433b-a2e8-2a1dd0066c97\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.416536 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.416859 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8k48v" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.425446 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.425553 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.425742 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.425855 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.426047 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c759802a-33e9-49e6-b764-7504e867562f-client-ca\") pod \"controller-manager-879f6c89f-9drmk\" (UID: \"c759802a-33e9-49e6-b764-7504e867562f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9drmk" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.426110 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c759802a-33e9-49e6-b764-7504e867562f-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9drmk\" (UID: \"c759802a-33e9-49e6-b764-7504e867562f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9drmk" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.426156 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/cc0f9f33-98c7-433b-a2e8-2a1dd0066c97-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hsdts\" (UID: \"cc0f9f33-98c7-433b-a2e8-2a1dd0066c97\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.426173 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc0f9f33-98c7-433b-a2e8-2a1dd0066c97-serving-cert\") pod \"apiserver-7bbb656c7d-hsdts\" (UID: \"cc0f9f33-98c7-433b-a2e8-2a1dd0066c97\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.426191 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.426209 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ff428ab-6c11-4571-9717-9ac0bfe462d4-serving-cert\") pod \"route-controller-manager-6576b87f9c-wfh87\" (UID: \"0ff428ab-6c11-4571-9717-9ac0bfe462d4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.426227 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/cc0f9f33-98c7-433b-a2e8-2a1dd0066c97-audit-policies\") pod \"apiserver-7bbb656c7d-hsdts\" (UID: \"cc0f9f33-98c7-433b-a2e8-2a1dd0066c97\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.426244 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/cc0f9f33-98c7-433b-a2e8-2a1dd0066c97-encryption-config\") pod \"apiserver-7bbb656c7d-hsdts\" (UID: \"cc0f9f33-98c7-433b-a2e8-2a1dd0066c97\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.426260 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c759802a-33e9-49e6-b764-7504e867562f-serving-cert\") pod \"controller-manager-879f6c89f-9drmk\" (UID: \"c759802a-33e9-49e6-b764-7504e867562f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9drmk" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.426285 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb4ba104-3024-4e76-a15c-77decd3a67c6-config\") pod \"machine-approver-56656f9798-l78hd\" (UID: \"bb4ba104-3024-4e76-a15c-77decd3a67c6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l78hd" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.426300 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.426320 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ced32e5c-2931-4a96-a779-1c1c6b71243e-audit-policies\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.426344 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jxg6\" (UniqueName: \"kubernetes.io/projected/c759802a-33e9-49e6-b764-7504e867562f-kube-api-access-4jxg6\") pod \"controller-manager-879f6c89f-9drmk\" (UID: \"c759802a-33e9-49e6-b764-7504e867562f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9drmk" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.426361 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0ff428ab-6c11-4571-9717-9ac0bfe462d4-client-ca\") pod \"route-controller-manager-6576b87f9c-wfh87\" (UID: \"0ff428ab-6c11-4571-9717-9ac0bfe462d4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.426377 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ced32e5c-2931-4a96-a779-1c1c6b71243e-audit-dir\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.426396 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.426413 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ff428ab-6c11-4571-9717-9ac0bfe462d4-config\") pod \"route-controller-manager-6576b87f9c-wfh87\" (UID: \"0ff428ab-6c11-4571-9717-9ac0bfe462d4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.426525 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.426697 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-dt2k4" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.427079 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.429219 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.429245 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvcdd\" (UniqueName: \"kubernetes.io/projected/0ff428ab-6c11-4571-9717-9ac0bfe462d4-kube-api-access-xvcdd\") pod \"route-controller-manager-6576b87f9c-wfh87\" (UID: \"0ff428ab-6c11-4571-9717-9ac0bfe462d4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.429263 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cc0f9f33-98c7-433b-a2e8-2a1dd0066c97-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hsdts\" (UID: \"cc0f9f33-98c7-433b-a2e8-2a1dd0066c97\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.429280 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.429295 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bb4ba104-3024-4e76-a15c-77decd3a67c6-auth-proxy-config\") pod \"machine-approver-56656f9798-l78hd\" (UID: \"bb4ba104-3024-4e76-a15c-77decd3a67c6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l78hd" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.429324 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.429344 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.429362 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.429385 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.429403 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cc0f9f33-98c7-433b-a2e8-2a1dd0066c97-audit-dir\") pod \"apiserver-7bbb656c7d-hsdts\" (UID: \"cc0f9f33-98c7-433b-a2e8-2a1dd0066c97\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.429424 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.429445 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hv9h\" (UniqueName: \"kubernetes.io/projected/bb4ba104-3024-4e76-a15c-77decd3a67c6-kube-api-access-5hv9h\") pod \"machine-approver-56656f9798-l78hd\" (UID: \"bb4ba104-3024-4e76-a15c-77decd3a67c6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l78hd" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.429463 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/bb4ba104-3024-4e76-a15c-77decd3a67c6-machine-approver-tls\") pod \"machine-approver-56656f9798-l78hd\" (UID: \"bb4ba104-3024-4e76-a15c-77decd3a67c6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l78hd" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.429478 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cc0f9f33-98c7-433b-a2e8-2a1dd0066c97-etcd-client\") pod \"apiserver-7bbb656c7d-hsdts\" (UID: \"cc0f9f33-98c7-433b-a2e8-2a1dd0066c97\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.429496 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c759802a-33e9-49e6-b764-7504e867562f-config\") pod \"controller-manager-879f6c89f-9drmk\" (UID: \"c759802a-33e9-49e6-b764-7504e867562f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9drmk" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.429522 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.429728 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.429956 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.430598 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.430811 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.430946 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-phxkp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.431029 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.431145 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.432352 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ddcbg"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.432537 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.433332 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.433377 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.433611 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.433789 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.436362 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-7bxtq"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.436403 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.436414 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.436578 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.436666 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ddcbg" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.436713 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2bh9c"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.437257 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2bh9c" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.437347 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.437425 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.437466 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.437522 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.437565 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.437523 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-7bxtq" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.437903 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.438572 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.438880 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.438978 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.439061 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.439176 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.439226 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.439315 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.439337 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-pcvlp"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.439383 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.439439 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.439634 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.439669 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.439797 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.439638 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.439932 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-pcvlp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.439982 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.440146 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.440094 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.440305 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.440440 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.440461 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-8nt7g"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.440565 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.441014 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-8nt7g" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.441806 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jhgsq"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.442372 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.442470 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zwd57"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.442755 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zwd57" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.444483 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.445133 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lxbgb"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.445438 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-xkp55"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.445831 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xkp55" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.445887 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lxbgb" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.446607 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.447408 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-jdvrw"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.448065 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jdvrw" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.449768 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rxt22"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.450421 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rxt22" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.450760 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-94zbw"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.451319 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-94zbw" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.451863 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-q9cfv"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.452276 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-q9cfv" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.529664 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.530483 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.531447 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.531495 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.531618 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.531624 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.531685 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532049 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532172 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c759802a-33e9-49e6-b764-7504e867562f-serving-cert\") pod \"controller-manager-879f6c89f-9drmk\" (UID: \"c759802a-33e9-49e6-b764-7504e867562f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9drmk" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532203 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb4ba104-3024-4e76-a15c-77decd3a67c6-config\") pod \"machine-approver-56656f9798-l78hd\" (UID: \"bb4ba104-3024-4e76-a15c-77decd3a67c6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l78hd" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532231 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532254 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ced32e5c-2931-4a96-a779-1c1c6b71243e-audit-policies\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532265 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532274 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jxg6\" (UniqueName: \"kubernetes.io/projected/c759802a-33e9-49e6-b764-7504e867562f-kube-api-access-4jxg6\") pod \"controller-manager-879f6c89f-9drmk\" (UID: \"c759802a-33e9-49e6-b764-7504e867562f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9drmk" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532296 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0ff428ab-6c11-4571-9717-9ac0bfe462d4-client-ca\") pod \"route-controller-manager-6576b87f9c-wfh87\" (UID: \"0ff428ab-6c11-4571-9717-9ac0bfe462d4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532318 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ced32e5c-2931-4a96-a779-1c1c6b71243e-audit-dir\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532339 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532361 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ff428ab-6c11-4571-9717-9ac0bfe462d4-config\") pod \"route-controller-manager-6576b87f9c-wfh87\" (UID: \"0ff428ab-6c11-4571-9717-9ac0bfe462d4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532386 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532407 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvcdd\" (UniqueName: \"kubernetes.io/projected/0ff428ab-6c11-4571-9717-9ac0bfe462d4-kube-api-access-xvcdd\") pod \"route-controller-manager-6576b87f9c-wfh87\" (UID: \"0ff428ab-6c11-4571-9717-9ac0bfe462d4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532429 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cc0f9f33-98c7-433b-a2e8-2a1dd0066c97-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hsdts\" (UID: \"cc0f9f33-98c7-433b-a2e8-2a1dd0066c97\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532453 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bb4ba104-3024-4e76-a15c-77decd3a67c6-auth-proxy-config\") pod \"machine-approver-56656f9798-l78hd\" (UID: \"bb4ba104-3024-4e76-a15c-77decd3a67c6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l78hd" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532476 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532501 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532525 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532547 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532572 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532598 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532622 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hv9h\" (UniqueName: \"kubernetes.io/projected/bb4ba104-3024-4e76-a15c-77decd3a67c6-kube-api-access-5hv9h\") pod \"machine-approver-56656f9798-l78hd\" (UID: \"bb4ba104-3024-4e76-a15c-77decd3a67c6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l78hd" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532644 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cc0f9f33-98c7-433b-a2e8-2a1dd0066c97-audit-dir\") pod \"apiserver-7bbb656c7d-hsdts\" (UID: \"cc0f9f33-98c7-433b-a2e8-2a1dd0066c97\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532672 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/bb4ba104-3024-4e76-a15c-77decd3a67c6-machine-approver-tls\") pod \"machine-approver-56656f9798-l78hd\" (UID: \"bb4ba104-3024-4e76-a15c-77decd3a67c6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l78hd" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532697 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cc0f9f33-98c7-433b-a2e8-2a1dd0066c97-etcd-client\") pod \"apiserver-7bbb656c7d-hsdts\" (UID: \"cc0f9f33-98c7-433b-a2e8-2a1dd0066c97\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532723 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c759802a-33e9-49e6-b764-7504e867562f-config\") pod \"controller-manager-879f6c89f-9drmk\" (UID: \"c759802a-33e9-49e6-b764-7504e867562f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9drmk" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532759 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532783 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpw6q\" (UniqueName: \"kubernetes.io/projected/ced32e5c-2931-4a96-a779-1c1c6b71243e-kube-api-access-tpw6q\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532836 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbwpg\" (UniqueName: \"kubernetes.io/projected/cc0f9f33-98c7-433b-a2e8-2a1dd0066c97-kube-api-access-nbwpg\") pod \"apiserver-7bbb656c7d-hsdts\" (UID: \"cc0f9f33-98c7-433b-a2e8-2a1dd0066c97\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532858 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c759802a-33e9-49e6-b764-7504e867562f-client-ca\") pod \"controller-manager-879f6c89f-9drmk\" (UID: \"c759802a-33e9-49e6-b764-7504e867562f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9drmk" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532880 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c759802a-33e9-49e6-b764-7504e867562f-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9drmk\" (UID: \"c759802a-33e9-49e6-b764-7504e867562f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9drmk" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532914 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/cc0f9f33-98c7-433b-a2e8-2a1dd0066c97-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hsdts\" (UID: \"cc0f9f33-98c7-433b-a2e8-2a1dd0066c97\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532937 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc0f9f33-98c7-433b-a2e8-2a1dd0066c97-serving-cert\") pod \"apiserver-7bbb656c7d-hsdts\" (UID: \"cc0f9f33-98c7-433b-a2e8-2a1dd0066c97\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532961 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532985 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/cc0f9f33-98c7-433b-a2e8-2a1dd0066c97-audit-policies\") pod \"apiserver-7bbb656c7d-hsdts\" (UID: \"cc0f9f33-98c7-433b-a2e8-2a1dd0066c97\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.533009 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/cc0f9f33-98c7-433b-a2e8-2a1dd0066c97-encryption-config\") pod \"apiserver-7bbb656c7d-hsdts\" (UID: \"cc0f9f33-98c7-433b-a2e8-2a1dd0066c97\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.533034 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ff428ab-6c11-4571-9717-9ac0bfe462d4-serving-cert\") pod \"route-controller-manager-6576b87f9c-wfh87\" (UID: \"0ff428ab-6c11-4571-9717-9ac0bfe462d4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.533620 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-s95w5"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.534478 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xwlzm"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.534693 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s95w5" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.535176 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-jj9rd"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.535650 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jj9rd" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.537680 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/cc0f9f33-98c7-433b-a2e8-2a1dd0066c97-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hsdts\" (UID: \"cc0f9f33-98c7-433b-a2e8-2a1dd0066c97\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.538369 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c759802a-33e9-49e6-b764-7504e867562f-client-ca\") pod \"controller-manager-879f6c89f-9drmk\" (UID: \"c759802a-33e9-49e6-b764-7504e867562f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9drmk" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.539038 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.539220 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c759802a-33e9-49e6-b764-7504e867562f-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9drmk\" (UID: \"c759802a-33e9-49e6-b764-7504e867562f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9drmk" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.539346 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xwlzm" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.540002 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/cc0f9f33-98c7-433b-a2e8-2a1dd0066c97-audit-policies\") pod \"apiserver-7bbb656c7d-hsdts\" (UID: \"cc0f9f33-98c7-433b-a2e8-2a1dd0066c97\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.542463 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.542892 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ced32e5c-2931-4a96-a779-1c1c6b71243e-audit-policies\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.543571 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0ff428ab-6c11-4571-9717-9ac0bfe462d4-client-ca\") pod \"route-controller-manager-6576b87f9c-wfh87\" (UID: \"0ff428ab-6c11-4571-9717-9ac0bfe462d4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.543607 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ced32e5c-2931-4a96-a779-1c1c6b71243e-audit-dir\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.532432 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.544736 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc0f9f33-98c7-433b-a2e8-2a1dd0066c97-serving-cert\") pod \"apiserver-7bbb656c7d-hsdts\" (UID: \"cc0f9f33-98c7-433b-a2e8-2a1dd0066c97\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.544870 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ff428ab-6c11-4571-9717-9ac0bfe462d4-config\") pod \"route-controller-manager-6576b87f9c-wfh87\" (UID: \"0ff428ab-6c11-4571-9717-9ac0bfe462d4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.545222 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.545960 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cc0f9f33-98c7-433b-a2e8-2a1dd0066c97-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hsdts\" (UID: \"cc0f9f33-98c7-433b-a2e8-2a1dd0066c97\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.546166 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bb4ba104-3024-4e76-a15c-77decd3a67c6-auth-proxy-config\") pod \"machine-approver-56656f9798-l78hd\" (UID: \"bb4ba104-3024-4e76-a15c-77decd3a67c6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l78hd" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.546564 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-5qkt6"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.546624 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cc0f9f33-98c7-433b-a2e8-2a1dd0066c97-audit-dir\") pod \"apiserver-7bbb656c7d-hsdts\" (UID: \"cc0f9f33-98c7-433b-a2e8-2a1dd0066c97\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.546732 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.547189 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-5qkt6" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.547390 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.547523 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c759802a-33e9-49e6-b764-7504e867562f-config\") pod \"controller-manager-879f6c89f-9drmk\" (UID: \"c759802a-33e9-49e6-b764-7504e867562f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9drmk" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.534249 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.534343 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.534394 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.547886 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb4ba104-3024-4e76-a15c-77decd3a67c6-config\") pod \"machine-approver-56656f9798-l78hd\" (UID: \"bb4ba104-3024-4e76-a15c-77decd3a67c6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l78hd" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.535019 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.535120 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.548295 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c759802a-33e9-49e6-b764-7504e867562f-serving-cert\") pod \"controller-manager-879f6c89f-9drmk\" (UID: \"c759802a-33e9-49e6-b764-7504e867562f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9drmk" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.548382 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.548428 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t8v4d"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.548722 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ff428ab-6c11-4571-9717-9ac0bfe462d4-serving-cert\") pod \"route-controller-manager-6576b87f9c-wfh87\" (UID: \"0ff428ab-6c11-4571-9717-9ac0bfe462d4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.548778 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.549038 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t8v4d" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.549061 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ls9ng"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.549490 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ls9ng" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.549510 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.549559 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/cc0f9f33-98c7-433b-a2e8-2a1dd0066c97-encryption-config\") pod \"apiserver-7bbb656c7d-hsdts\" (UID: \"cc0f9f33-98c7-433b-a2e8-2a1dd0066c97\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.549909 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gfnhv"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.550535 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gfnhv" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.550961 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-529ml"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.552546 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/bb4ba104-3024-4e76-a15c-77decd3a67c6-machine-approver-tls\") pod \"machine-approver-56656f9798-l78hd\" (UID: \"bb4ba104-3024-4e76-a15c-77decd3a67c6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l78hd" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.552555 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.552835 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z54z"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.553009 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-529ml" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.553267 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-v5pq6"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.555960 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z54z" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.556382 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330745-wbwkx"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.560510 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.560539 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-67jl7"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.560549 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9drmk"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.560558 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zg2sm"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.560639 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330745-wbwkx" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.556542 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-v5pq6" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.562302 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.565825 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.567062 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cc0f9f33-98c7-433b-a2e8-2a1dd0066c97-etcd-client\") pod \"apiserver-7bbb656c7d-hsdts\" (UID: \"cc0f9f33-98c7-433b-a2e8-2a1dd0066c97\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.567558 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8k48v"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.567599 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-82gdf"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.567666 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.570152 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.570734 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.573815 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-tv7sf"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.574697 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ddcbg"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.574760 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-rfcqs"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.576075 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-rfcqs" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.580226 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-94zbw"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.580260 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mx6p7"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.580270 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.584296 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.591711 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.592709 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.594884 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-xkp55"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.595435 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.605349 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t8v4d"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.606319 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-cw2cl"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.607326 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-529ml"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.608645 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-dt2k4"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.609984 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.611859 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lxbgb"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.617262 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-phxkp"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.618600 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jhgsq"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.619921 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qr2pg"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.620985 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2bh9c"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.621974 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-7cvhs"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.622826 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-7cvhs" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.624038 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-q9cfv"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.625074 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-pcvlp"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.626224 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-7bxtq"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.627147 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330745-wbwkx"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.628212 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xwlzm"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.629125 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.629363 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-jj9rd"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.630251 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-5qkt6"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.631589 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-jdvrw"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.632624 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-rfcqs"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.633648 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-fggsp"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.633955 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-console-serving-cert\") pod \"console-f9d7485db-pcvlp\" (UID: \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\") " pod="openshift-console/console-f9d7485db-pcvlp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.633981 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd2f3a74-d8f7-4970-b748-39036ccd4aab-serving-cert\") pod \"etcd-operator-b45778765-qr2pg\" (UID: \"dd2f3a74-d8f7-4970-b748-39036ccd4aab\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qr2pg" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.634031 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kxfq\" (UniqueName: \"kubernetes.io/projected/5d295e40-2e2f-419e-9db4-cee7392d913e-kube-api-access-8kxfq\") pod \"control-plane-machine-set-operator-78cbb6b69f-8k48v\" (UID: \"5d295e40-2e2f-419e-9db4-cee7392d913e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8k48v" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.634067 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/dd2f3a74-d8f7-4970-b748-39036ccd4aab-etcd-service-ca\") pod \"etcd-operator-b45778765-qr2pg\" (UID: \"dd2f3a74-d8f7-4970-b748-39036ccd4aab\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qr2pg" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.634088 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv6qg\" (UniqueName: \"kubernetes.io/projected/67543fb5-5742-4067-a953-7d94ac45e415-kube-api-access-bv6qg\") pod \"router-default-5444994796-8nt7g\" (UID: \"67543fb5-5742-4067-a953-7d94ac45e415\") " pod="openshift-ingress/router-default-5444994796-8nt7g" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.634146 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-service-ca\") pod \"console-f9d7485db-pcvlp\" (UID: \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\") " pod="openshift-console/console-f9d7485db-pcvlp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.634169 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/dd2f3a74-d8f7-4970-b748-39036ccd4aab-etcd-ca\") pod \"etcd-operator-b45778765-qr2pg\" (UID: \"dd2f3a74-d8f7-4970-b748-39036ccd4aab\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qr2pg" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.634216 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-console-config\") pod \"console-f9d7485db-pcvlp\" (UID: \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\") " pod="openshift-console/console-f9d7485db-pcvlp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.634234 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/dd2f3a74-d8f7-4970-b748-39036ccd4aab-etcd-client\") pod \"etcd-operator-b45778765-qr2pg\" (UID: \"dd2f3a74-d8f7-4970-b748-39036ccd4aab\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qr2pg" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.634252 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/67543fb5-5742-4067-a953-7d94ac45e415-stats-auth\") pod \"router-default-5444994796-8nt7g\" (UID: \"67543fb5-5742-4067-a953-7d94ac45e415\") " pod="openshift-ingress/router-default-5444994796-8nt7g" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.634269 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/67543fb5-5742-4067-a953-7d94ac45e415-metrics-certs\") pod \"router-default-5444994796-8nt7g\" (UID: \"67543fb5-5742-4067-a953-7d94ac45e415\") " pod="openshift-ingress/router-default-5444994796-8nt7g" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.634289 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngjg8\" (UniqueName: \"kubernetes.io/projected/2e645a3d-0c4a-4e27-83ff-5bb01f1da515-kube-api-access-ngjg8\") pod \"console-operator-58897d9998-phxkp\" (UID: \"2e645a3d-0c4a-4e27-83ff-5bb01f1da515\") " pod="openshift-console-operator/console-operator-58897d9998-phxkp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.634303 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd2f3a74-d8f7-4970-b748-39036ccd4aab-config\") pod \"etcd-operator-b45778765-qr2pg\" (UID: \"dd2f3a74-d8f7-4970-b748-39036ccd4aab\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qr2pg" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.634331 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e645a3d-0c4a-4e27-83ff-5bb01f1da515-config\") pod \"console-operator-58897d9998-phxkp\" (UID: \"2e645a3d-0c4a-4e27-83ff-5bb01f1da515\") " pod="openshift-console-operator/console-operator-58897d9998-phxkp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.634349 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-trusted-ca-bundle\") pod \"console-f9d7485db-pcvlp\" (UID: \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\") " pod="openshift-console/console-f9d7485db-pcvlp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.634365 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/67543fb5-5742-4067-a953-7d94ac45e415-service-ca-bundle\") pod \"router-default-5444994796-8nt7g\" (UID: \"67543fb5-5742-4067-a953-7d94ac45e415\") " pod="openshift-ingress/router-default-5444994796-8nt7g" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.634382 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwnzc\" (UniqueName: \"kubernetes.io/projected/6158d485-779d-4fe3-8772-32b33a67b1f0-kube-api-access-fwnzc\") pod \"migrator-59844c95c7-jdvrw\" (UID: \"6158d485-779d-4fe3-8772-32b33a67b1f0\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jdvrw" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.634400 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2e645a3d-0c4a-4e27-83ff-5bb01f1da515-trusted-ca\") pod \"console-operator-58897d9998-phxkp\" (UID: \"2e645a3d-0c4a-4e27-83ff-5bb01f1da515\") " pod="openshift-console-operator/console-operator-58897d9998-phxkp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.634418 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhghz\" (UniqueName: \"kubernetes.io/projected/dd2f3a74-d8f7-4970-b748-39036ccd4aab-kube-api-access-vhghz\") pod \"etcd-operator-b45778765-qr2pg\" (UID: \"dd2f3a74-d8f7-4970-b748-39036ccd4aab\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qr2pg" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.634437 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/5d295e40-2e2f-419e-9db4-cee7392d913e-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-8k48v\" (UID: \"5d295e40-2e2f-419e-9db4-cee7392d913e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8k48v" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.634454 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxszd\" (UniqueName: \"kubernetes.io/projected/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-kube-api-access-qxszd\") pod \"console-f9d7485db-pcvlp\" (UID: \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\") " pod="openshift-console/console-f9d7485db-pcvlp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.634472 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/67543fb5-5742-4067-a953-7d94ac45e415-default-certificate\") pod \"router-default-5444994796-8nt7g\" (UID: \"67543fb5-5742-4067-a953-7d94ac45e415\") " pod="openshift-ingress/router-default-5444994796-8nt7g" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.634498 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2e645a3d-0c4a-4e27-83ff-5bb01f1da515-serving-cert\") pod \"console-operator-58897d9998-phxkp\" (UID: \"2e645a3d-0c4a-4e27-83ff-5bb01f1da515\") " pod="openshift-console-operator/console-operator-58897d9998-phxkp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.634514 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-oauth-serving-cert\") pod \"console-f9d7485db-pcvlp\" (UID: \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\") " pod="openshift-console/console-f9d7485db-pcvlp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.634531 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-console-oauth-config\") pod \"console-f9d7485db-pcvlp\" (UID: \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\") " pod="openshift-console/console-f9d7485db-pcvlp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.634825 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-fggsp"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.635072 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-fggsp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.635971 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-s95w5"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.637554 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z54z"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.638070 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zwd57"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.639129 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rxt22"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.640204 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ls9ng"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.641158 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-v5pq6"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.642120 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gfnhv"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.643259 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-7cvhs"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.644652 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-xn2j9"] Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.645239 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-xn2j9" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.649151 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.676076 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.689114 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.709187 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.731045 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.735177 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/dd2f3a74-d8f7-4970-b748-39036ccd4aab-etcd-ca\") pod \"etcd-operator-b45778765-qr2pg\" (UID: \"dd2f3a74-d8f7-4970-b748-39036ccd4aab\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qr2pg" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.735237 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/67543fb5-5742-4067-a953-7d94ac45e415-stats-auth\") pod \"router-default-5444994796-8nt7g\" (UID: \"67543fb5-5742-4067-a953-7d94ac45e415\") " pod="openshift-ingress/router-default-5444994796-8nt7g" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.735255 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/67543fb5-5742-4067-a953-7d94ac45e415-metrics-certs\") pod \"router-default-5444994796-8nt7g\" (UID: \"67543fb5-5742-4067-a953-7d94ac45e415\") " pod="openshift-ingress/router-default-5444994796-8nt7g" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.735276 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-console-config\") pod \"console-f9d7485db-pcvlp\" (UID: \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\") " pod="openshift-console/console-f9d7485db-pcvlp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.735290 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/dd2f3a74-d8f7-4970-b748-39036ccd4aab-etcd-client\") pod \"etcd-operator-b45778765-qr2pg\" (UID: \"dd2f3a74-d8f7-4970-b748-39036ccd4aab\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qr2pg" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.735310 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngjg8\" (UniqueName: \"kubernetes.io/projected/2e645a3d-0c4a-4e27-83ff-5bb01f1da515-kube-api-access-ngjg8\") pod \"console-operator-58897d9998-phxkp\" (UID: \"2e645a3d-0c4a-4e27-83ff-5bb01f1da515\") " pod="openshift-console-operator/console-operator-58897d9998-phxkp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.735327 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd2f3a74-d8f7-4970-b748-39036ccd4aab-config\") pod \"etcd-operator-b45778765-qr2pg\" (UID: \"dd2f3a74-d8f7-4970-b748-39036ccd4aab\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qr2pg" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.735352 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e645a3d-0c4a-4e27-83ff-5bb01f1da515-config\") pod \"console-operator-58897d9998-phxkp\" (UID: \"2e645a3d-0c4a-4e27-83ff-5bb01f1da515\") " pod="openshift-console-operator/console-operator-58897d9998-phxkp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.735368 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/67543fb5-5742-4067-a953-7d94ac45e415-service-ca-bundle\") pod \"router-default-5444994796-8nt7g\" (UID: \"67543fb5-5742-4067-a953-7d94ac45e415\") " pod="openshift-ingress/router-default-5444994796-8nt7g" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.735385 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-trusted-ca-bundle\") pod \"console-f9d7485db-pcvlp\" (UID: \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\") " pod="openshift-console/console-f9d7485db-pcvlp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.735400 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwnzc\" (UniqueName: \"kubernetes.io/projected/6158d485-779d-4fe3-8772-32b33a67b1f0-kube-api-access-fwnzc\") pod \"migrator-59844c95c7-jdvrw\" (UID: \"6158d485-779d-4fe3-8772-32b33a67b1f0\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jdvrw" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.735413 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2e645a3d-0c4a-4e27-83ff-5bb01f1da515-trusted-ca\") pod \"console-operator-58897d9998-phxkp\" (UID: \"2e645a3d-0c4a-4e27-83ff-5bb01f1da515\") " pod="openshift-console-operator/console-operator-58897d9998-phxkp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.735428 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhghz\" (UniqueName: \"kubernetes.io/projected/dd2f3a74-d8f7-4970-b748-39036ccd4aab-kube-api-access-vhghz\") pod \"etcd-operator-b45778765-qr2pg\" (UID: \"dd2f3a74-d8f7-4970-b748-39036ccd4aab\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qr2pg" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.735444 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxszd\" (UniqueName: \"kubernetes.io/projected/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-kube-api-access-qxszd\") pod \"console-f9d7485db-pcvlp\" (UID: \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\") " pod="openshift-console/console-f9d7485db-pcvlp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.735459 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/67543fb5-5742-4067-a953-7d94ac45e415-default-certificate\") pod \"router-default-5444994796-8nt7g\" (UID: \"67543fb5-5742-4067-a953-7d94ac45e415\") " pod="openshift-ingress/router-default-5444994796-8nt7g" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.735478 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/5d295e40-2e2f-419e-9db4-cee7392d913e-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-8k48v\" (UID: \"5d295e40-2e2f-419e-9db4-cee7392d913e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8k48v" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.735500 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2e645a3d-0c4a-4e27-83ff-5bb01f1da515-serving-cert\") pod \"console-operator-58897d9998-phxkp\" (UID: \"2e645a3d-0c4a-4e27-83ff-5bb01f1da515\") " pod="openshift-console-operator/console-operator-58897d9998-phxkp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.735514 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-oauth-serving-cert\") pod \"console-f9d7485db-pcvlp\" (UID: \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\") " pod="openshift-console/console-f9d7485db-pcvlp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.735530 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-console-oauth-config\") pod \"console-f9d7485db-pcvlp\" (UID: \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\") " pod="openshift-console/console-f9d7485db-pcvlp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.735556 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd2f3a74-d8f7-4970-b748-39036ccd4aab-serving-cert\") pod \"etcd-operator-b45778765-qr2pg\" (UID: \"dd2f3a74-d8f7-4970-b748-39036ccd4aab\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qr2pg" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.735572 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-console-serving-cert\") pod \"console-f9d7485db-pcvlp\" (UID: \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\") " pod="openshift-console/console-f9d7485db-pcvlp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.735590 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kxfq\" (UniqueName: \"kubernetes.io/projected/5d295e40-2e2f-419e-9db4-cee7392d913e-kube-api-access-8kxfq\") pod \"control-plane-machine-set-operator-78cbb6b69f-8k48v\" (UID: \"5d295e40-2e2f-419e-9db4-cee7392d913e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8k48v" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.735614 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv6qg\" (UniqueName: \"kubernetes.io/projected/67543fb5-5742-4067-a953-7d94ac45e415-kube-api-access-bv6qg\") pod \"router-default-5444994796-8nt7g\" (UID: \"67543fb5-5742-4067-a953-7d94ac45e415\") " pod="openshift-ingress/router-default-5444994796-8nt7g" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.735637 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/dd2f3a74-d8f7-4970-b748-39036ccd4aab-etcd-service-ca\") pod \"etcd-operator-b45778765-qr2pg\" (UID: \"dd2f3a74-d8f7-4970-b748-39036ccd4aab\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qr2pg" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.735660 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-service-ca\") pod \"console-f9d7485db-pcvlp\" (UID: \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\") " pod="openshift-console/console-f9d7485db-pcvlp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.736030 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-console-config\") pod \"console-f9d7485db-pcvlp\" (UID: \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\") " pod="openshift-console/console-f9d7485db-pcvlp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.736301 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-service-ca\") pod \"console-f9d7485db-pcvlp\" (UID: \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\") " pod="openshift-console/console-f9d7485db-pcvlp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.736874 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-trusted-ca-bundle\") pod \"console-f9d7485db-pcvlp\" (UID: \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\") " pod="openshift-console/console-f9d7485db-pcvlp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.736920 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e645a3d-0c4a-4e27-83ff-5bb01f1da515-config\") pod \"console-operator-58897d9998-phxkp\" (UID: \"2e645a3d-0c4a-4e27-83ff-5bb01f1da515\") " pod="openshift-console-operator/console-operator-58897d9998-phxkp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.736941 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/dd2f3a74-d8f7-4970-b748-39036ccd4aab-etcd-ca\") pod \"etcd-operator-b45778765-qr2pg\" (UID: \"dd2f3a74-d8f7-4970-b748-39036ccd4aab\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qr2pg" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.737082 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd2f3a74-d8f7-4970-b748-39036ccd4aab-config\") pod \"etcd-operator-b45778765-qr2pg\" (UID: \"dd2f3a74-d8f7-4970-b748-39036ccd4aab\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qr2pg" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.737311 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2e645a3d-0c4a-4e27-83ff-5bb01f1da515-trusted-ca\") pod \"console-operator-58897d9998-phxkp\" (UID: \"2e645a3d-0c4a-4e27-83ff-5bb01f1da515\") " pod="openshift-console-operator/console-operator-58897d9998-phxkp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.737858 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/dd2f3a74-d8f7-4970-b748-39036ccd4aab-etcd-client\") pod \"etcd-operator-b45778765-qr2pg\" (UID: \"dd2f3a74-d8f7-4970-b748-39036ccd4aab\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qr2pg" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.739072 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-console-oauth-config\") pod \"console-f9d7485db-pcvlp\" (UID: \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\") " pod="openshift-console/console-f9d7485db-pcvlp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.739305 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-oauth-serving-cert\") pod \"console-f9d7485db-pcvlp\" (UID: \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\") " pod="openshift-console/console-f9d7485db-pcvlp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.739365 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/dd2f3a74-d8f7-4970-b748-39036ccd4aab-etcd-service-ca\") pod \"etcd-operator-b45778765-qr2pg\" (UID: \"dd2f3a74-d8f7-4970-b748-39036ccd4aab\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qr2pg" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.739796 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/67543fb5-5742-4067-a953-7d94ac45e415-default-certificate\") pod \"router-default-5444994796-8nt7g\" (UID: \"67543fb5-5742-4067-a953-7d94ac45e415\") " pod="openshift-ingress/router-default-5444994796-8nt7g" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.740324 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/5d295e40-2e2f-419e-9db4-cee7392d913e-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-8k48v\" (UID: \"5d295e40-2e2f-419e-9db4-cee7392d913e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8k48v" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.740428 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-console-serving-cert\") pod \"console-f9d7485db-pcvlp\" (UID: \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\") " pod="openshift-console/console-f9d7485db-pcvlp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.741717 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2e645a3d-0c4a-4e27-83ff-5bb01f1da515-serving-cert\") pod \"console-operator-58897d9998-phxkp\" (UID: \"2e645a3d-0c4a-4e27-83ff-5bb01f1da515\") " pod="openshift-console-operator/console-operator-58897d9998-phxkp" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.742854 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd2f3a74-d8f7-4970-b748-39036ccd4aab-serving-cert\") pod \"etcd-operator-b45778765-qr2pg\" (UID: \"dd2f3a74-d8f7-4970-b748-39036ccd4aab\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qr2pg" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.750632 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.759070 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/67543fb5-5742-4067-a953-7d94ac45e415-stats-auth\") pod \"router-default-5444994796-8nt7g\" (UID: \"67543fb5-5742-4067-a953-7d94ac45e415\") " pod="openshift-ingress/router-default-5444994796-8nt7g" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.769882 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.778807 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/67543fb5-5742-4067-a953-7d94ac45e415-metrics-certs\") pod \"router-default-5444994796-8nt7g\" (UID: \"67543fb5-5742-4067-a953-7d94ac45e415\") " pod="openshift-ingress/router-default-5444994796-8nt7g" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.790654 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.809174 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.816940 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/67543fb5-5742-4067-a953-7d94ac45e415-service-ca-bundle\") pod \"router-default-5444994796-8nt7g\" (UID: \"67543fb5-5742-4067-a953-7d94ac45e415\") " pod="openshift-ingress/router-default-5444994796-8nt7g" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.838342 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.849074 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.869671 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.890014 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.909871 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.929721 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.950450 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.970265 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 07 13:47:37 crc kubenswrapper[4959]: I1007 13:47:37.989435 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.009657 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.030301 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.050157 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.069531 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.089987 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.109887 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.129729 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.150349 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.170430 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.190624 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.210733 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.229527 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.250037 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.270414 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.289676 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.309571 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.329568 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.350548 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.370354 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.389553 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.410030 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.430153 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.451066 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.468137 4959 request.go:700] Waited for 1.015541408s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-service-ca-operator/configmaps?fieldSelector=metadata.name%3Dservice-ca-operator-config&limit=500&resourceVersion=0 Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.470599 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.491145 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.550796 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.574757 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.590367 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.637902 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpw6q\" (UniqueName: \"kubernetes.io/projected/ced32e5c-2931-4a96-a779-1c1c6b71243e-kube-api-access-tpw6q\") pod \"oauth-openshift-558db77b4-67jl7\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.649144 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbwpg\" (UniqueName: \"kubernetes.io/projected/cc0f9f33-98c7-433b-a2e8-2a1dd0066c97-kube-api-access-nbwpg\") pod \"apiserver-7bbb656c7d-hsdts\" (UID: \"cc0f9f33-98c7-433b-a2e8-2a1dd0066c97\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.649493 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.670194 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.690178 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.709730 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.729971 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.734687 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.759271 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.769897 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.799700 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.826926 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jxg6\" (UniqueName: \"kubernetes.io/projected/c759802a-33e9-49e6-b764-7504e867562f-kube-api-access-4jxg6\") pod \"controller-manager-879f6c89f-9drmk\" (UID: \"c759802a-33e9-49e6-b764-7504e867562f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9drmk" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.846159 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hv9h\" (UniqueName: \"kubernetes.io/projected/bb4ba104-3024-4e76-a15c-77decd3a67c6-kube-api-access-5hv9h\") pod \"machine-approver-56656f9798-l78hd\" (UID: \"bb4ba104-3024-4e76-a15c-77decd3a67c6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l78hd" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.849769 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.869713 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.893212 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9drmk" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.910153 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.912527 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvcdd\" (UniqueName: \"kubernetes.io/projected/0ff428ab-6c11-4571-9717-9ac0bfe462d4-kube-api-access-xvcdd\") pod \"route-controller-manager-6576b87f9c-wfh87\" (UID: \"0ff428ab-6c11-4571-9717-9ac0bfe462d4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.931298 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.941830 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.950486 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.970589 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.985483 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-67jl7"] Oct 07 13:47:38 crc kubenswrapper[4959]: I1007 13:47:38.990527 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 07 13:47:39 crc kubenswrapper[4959]: W1007 13:47:39.001312 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podced32e5c_2931_4a96_a779_1c1c6b71243e.slice/crio-69b3d8090a7773777e3906944bb7694b711e47ef8f695739356b03ee05af5146 WatchSource:0}: Error finding container 69b3d8090a7773777e3906944bb7694b711e47ef8f695739356b03ee05af5146: Status 404 returned error can't find the container with id 69b3d8090a7773777e3906944bb7694b711e47ef8f695739356b03ee05af5146 Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.009338 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.033318 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.049966 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.069770 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.095081 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l78hd" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.096379 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.096929 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9drmk"] Oct 07 13:47:39 crc kubenswrapper[4959]: W1007 13:47:39.106540 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc759802a_33e9_49e6_b764_7504e867562f.slice/crio-3f94177413fca52e34dc650f9cab31a0b80c13a88f4fbc3dd47b6024fec3f45e WatchSource:0}: Error finding container 3f94177413fca52e34dc650f9cab31a0b80c13a88f4fbc3dd47b6024fec3f45e: Status 404 returned error can't find the container with id 3f94177413fca52e34dc650f9cab31a0b80c13a88f4fbc3dd47b6024fec3f45e Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.109245 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 07 13:47:39 crc kubenswrapper[4959]: W1007 13:47:39.114842 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb4ba104_3024_4e76_a15c_77decd3a67c6.slice/crio-654b585d9c54e3b2261f568c2049e1346372b2d177ce0bf6a9cc41ff5d2662fe WatchSource:0}: Error finding container 654b585d9c54e3b2261f568c2049e1346372b2d177ce0bf6a9cc41ff5d2662fe: Status 404 returned error can't find the container with id 654b585d9c54e3b2261f568c2049e1346372b2d177ce0bf6a9cc41ff5d2662fe Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.124200 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts"] Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.129867 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 07 13:47:39 crc kubenswrapper[4959]: W1007 13:47:39.138039 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc0f9f33_98c7_433b_a2e8_2a1dd0066c97.slice/crio-5590c476791716dadb05160b0f79c5ee3a278a346feaf938f5c200b221823a84 WatchSource:0}: Error finding container 5590c476791716dadb05160b0f79c5ee3a278a346feaf938f5c200b221823a84: Status 404 returned error can't find the container with id 5590c476791716dadb05160b0f79c5ee3a278a346feaf938f5c200b221823a84 Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.150095 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.170472 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.174335 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.189660 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.211358 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.229749 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.253713 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.270566 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.291018 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.314546 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.329780 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.350794 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.355796 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87"] Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.369776 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.390617 4959 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.395436 4959 generic.go:334] "Generic (PLEG): container finished" podID="cc0f9f33-98c7-433b-a2e8-2a1dd0066c97" containerID="d41dc703935db525a0002273dd14255d8cf70f8700a04fc6a5389125c7d4fe44" exitCode=0 Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.395508 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" event={"ID":"cc0f9f33-98c7-433b-a2e8-2a1dd0066c97","Type":"ContainerDied","Data":"d41dc703935db525a0002273dd14255d8cf70f8700a04fc6a5389125c7d4fe44"} Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.395572 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" event={"ID":"cc0f9f33-98c7-433b-a2e8-2a1dd0066c97","Type":"ContainerStarted","Data":"5590c476791716dadb05160b0f79c5ee3a278a346feaf938f5c200b221823a84"} Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.398667 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" event={"ID":"ced32e5c-2931-4a96-a779-1c1c6b71243e","Type":"ContainerStarted","Data":"91579ad9e4cf546e01320e6a4a2badc05b6ad3da8ae5dde2d16d52a509cbfdfe"} Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.398714 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" event={"ID":"ced32e5c-2931-4a96-a779-1c1c6b71243e","Type":"ContainerStarted","Data":"69b3d8090a7773777e3906944bb7694b711e47ef8f695739356b03ee05af5146"} Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.398940 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.401376 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9drmk" event={"ID":"c759802a-33e9-49e6-b764-7504e867562f","Type":"ContainerStarted","Data":"0053e3376bc6e7a8a2a6accdb09ba29008fd73067a965a6b29ce48e43a69f3df"} Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.401405 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9drmk" event={"ID":"c759802a-33e9-49e6-b764-7504e867562f","Type":"ContainerStarted","Data":"3f94177413fca52e34dc650f9cab31a0b80c13a88f4fbc3dd47b6024fec3f45e"} Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.401587 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-9drmk" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.402598 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87" event={"ID":"0ff428ab-6c11-4571-9717-9ac0bfe462d4","Type":"ContainerStarted","Data":"0db27f84e0fc2074ebc40d626583ccda52be1b4d22894a57733d49ebe30d3816"} Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.402662 4959 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-67jl7 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.8:6443/healthz\": dial tcp 10.217.0.8:6443: connect: connection refused" start-of-body= Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.402696 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" podUID="ced32e5c-2931-4a96-a779-1c1c6b71243e" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.8:6443/healthz\": dial tcp 10.217.0.8:6443: connect: connection refused" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.404208 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l78hd" event={"ID":"bb4ba104-3024-4e76-a15c-77decd3a67c6","Type":"ContainerStarted","Data":"b0aac16d55e64fd3aeea7427c0c56db4f07f8c01c02932398808aabc38394730"} Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.404229 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l78hd" event={"ID":"bb4ba104-3024-4e76-a15c-77decd3a67c6","Type":"ContainerStarted","Data":"654b585d9c54e3b2261f568c2049e1346372b2d177ce0bf6a9cc41ff5d2662fe"} Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.405235 4959 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-9drmk container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.405279 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-9drmk" podUID="c759802a-33e9-49e6-b764-7504e867562f" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.410219 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.429364 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.450234 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.470088 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.487868 4959 request.go:700] Waited for 1.842341663s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dnode-bootstrapper-token&limit=500&resourceVersion=0 Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.489960 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.528920 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxszd\" (UniqueName: \"kubernetes.io/projected/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-kube-api-access-qxszd\") pod \"console-f9d7485db-pcvlp\" (UID: \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\") " pod="openshift-console/console-f9d7485db-pcvlp" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.549309 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngjg8\" (UniqueName: \"kubernetes.io/projected/2e645a3d-0c4a-4e27-83ff-5bb01f1da515-kube-api-access-ngjg8\") pod \"console-operator-58897d9998-phxkp\" (UID: \"2e645a3d-0c4a-4e27-83ff-5bb01f1da515\") " pod="openshift-console-operator/console-operator-58897d9998-phxkp" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.566653 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhghz\" (UniqueName: \"kubernetes.io/projected/dd2f3a74-d8f7-4970-b748-39036ccd4aab-kube-api-access-vhghz\") pod \"etcd-operator-b45778765-qr2pg\" (UID: \"dd2f3a74-d8f7-4970-b748-39036ccd4aab\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qr2pg" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.585447 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwnzc\" (UniqueName: \"kubernetes.io/projected/6158d485-779d-4fe3-8772-32b33a67b1f0-kube-api-access-fwnzc\") pod \"migrator-59844c95c7-jdvrw\" (UID: \"6158d485-779d-4fe3-8772-32b33a67b1f0\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jdvrw" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.607761 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kxfq\" (UniqueName: \"kubernetes.io/projected/5d295e40-2e2f-419e-9db4-cee7392d913e-kube-api-access-8kxfq\") pod \"control-plane-machine-set-operator-78cbb6b69f-8k48v\" (UID: \"5d295e40-2e2f-419e-9db4-cee7392d913e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8k48v" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.625615 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv6qg\" (UniqueName: \"kubernetes.io/projected/67543fb5-5742-4067-a953-7d94ac45e415-kube-api-access-bv6qg\") pod \"router-default-5444994796-8nt7g\" (UID: \"67543fb5-5742-4067-a953-7d94ac45e415\") " pod="openshift-ingress/router-default-5444994796-8nt7g" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.657984 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrbcf\" (UniqueName: \"kubernetes.io/projected/dcc8e980-62b6-48fb-8339-fab83032bd0a-kube-api-access-qrbcf\") pod \"authentication-operator-69f744f599-82gdf\" (UID: \"dcc8e980-62b6-48fb-8339-fab83032bd0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-82gdf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.658048 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/fe29debd-926a-4fd7-900c-b91a4d203436-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-2bh9c\" (UID: \"fe29debd-926a-4fd7-900c-b91a4d203436\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2bh9c" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.658073 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/de5396bf-4d94-46be-87e1-7b587147a6a4-auth-proxy-config\") pod \"machine-config-operator-74547568cd-s95w5\" (UID: \"de5396bf-4d94-46be-87e1-7b587147a6a4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s95w5" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.658146 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsm8h\" (UniqueName: \"kubernetes.io/projected/24ccfcc4-25a6-4264-86b1-6ba0097b2614-kube-api-access-xsm8h\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.658260 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f25b8da6-8185-4faa-b800-3deef3a9b06b-serving-cert\") pod \"openshift-config-operator-7777fb866f-xkp55\" (UID: \"f25b8da6-8185-4faa-b800-3deef3a9b06b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xkp55" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.658404 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9872926d-9b2f-489a-804d-749ec1e2c136-serving-cert\") pod \"service-ca-operator-777779d784-q9cfv\" (UID: \"9872926d-9b2f-489a-804d-749ec1e2c136\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-q9cfv" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.658497 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k25t2\" (UniqueName: \"kubernetes.io/projected/4a2a7f86-33ef-42d0-93b2-765b74d5c629-kube-api-access-k25t2\") pod \"openshift-controller-manager-operator-756b6f6bc6-mx6p7\" (UID: \"4a2a7f86-33ef-42d0-93b2-765b74d5c629\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mx6p7" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.658557 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b748d317-5cb9-4fd3-a6bc-4a83d2f77d97-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lxbgb\" (UID: \"b748d317-5cb9-4fd3-a6bc-4a83d2f77d97\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lxbgb" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.658676 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmk25\" (UniqueName: \"kubernetes.io/projected/fe29debd-926a-4fd7-900c-b91a4d203436-kube-api-access-jmk25\") pod \"cluster-samples-operator-665b6dd947-2bh9c\" (UID: \"fe29debd-926a-4fd7-900c-b91a4d203436\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2bh9c" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.659034 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/24ccfcc4-25a6-4264-86b1-6ba0097b2614-installation-pull-secrets\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.659245 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47a07011-2fae-40a2-9fb1-fec3461d2352-config\") pod \"kube-controller-manager-operator-78b949d7b-ddcbg\" (UID: \"47a07011-2fae-40a2-9fb1-fec3461d2352\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ddcbg" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.659309 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b816c481-6891-479c-99db-95f6e4abbc05-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-94zbw\" (UID: \"b816c481-6891-479c-99db-95f6e4abbc05\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-94zbw" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.659417 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/24ccfcc4-25a6-4264-86b1-6ba0097b2614-bound-sa-token\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.659450 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1d9c7700-8ba2-444e-a3f0-5c5c6dd05585-images\") pod \"machine-api-operator-5694c8668f-dt2k4\" (UID: \"1d9c7700-8ba2-444e-a3f0-5c5c6dd05585\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dt2k4" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.659487 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cl266\" (UniqueName: \"kubernetes.io/projected/9872926d-9b2f-489a-804d-749ec1e2c136-kube-api-access-cl266\") pod \"service-ca-operator-777779d784-q9cfv\" (UID: \"9872926d-9b2f-489a-804d-749ec1e2c136\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-q9cfv" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.659541 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/33deb07b-91b4-45f1-88d4-deed38a97f36-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-zwd57\" (UID: \"33deb07b-91b4-45f1-88d4-deed38a97f36\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zwd57" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.659617 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/33deb07b-91b4-45f1-88d4-deed38a97f36-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-zwd57\" (UID: \"33deb07b-91b4-45f1-88d4-deed38a97f36\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zwd57" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.659668 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/1d9c7700-8ba2-444e-a3f0-5c5c6dd05585-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-dt2k4\" (UID: \"1d9c7700-8ba2-444e-a3f0-5c5c6dd05585\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dt2k4" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.659712 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9dwg\" (UniqueName: \"kubernetes.io/projected/1d9c7700-8ba2-444e-a3f0-5c5c6dd05585-kube-api-access-d9dwg\") pod \"machine-api-operator-5694c8668f-dt2k4\" (UID: \"1d9c7700-8ba2-444e-a3f0-5c5c6dd05585\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dt2k4" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.659764 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9872926d-9b2f-489a-804d-749ec1e2c136-config\") pod \"service-ca-operator-777779d784-q9cfv\" (UID: \"9872926d-9b2f-489a-804d-749ec1e2c136\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-q9cfv" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.659813 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/24ccfcc4-25a6-4264-86b1-6ba0097b2614-registry-tls\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.659856 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/de5396bf-4d94-46be-87e1-7b587147a6a4-images\") pod \"machine-config-operator-74547568cd-s95w5\" (UID: \"de5396bf-4d94-46be-87e1-7b587147a6a4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s95w5" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.659906 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/24ccfcc4-25a6-4264-86b1-6ba0097b2614-ca-trust-extracted\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.659991 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/de5396bf-4d94-46be-87e1-7b587147a6a4-proxy-tls\") pod \"machine-config-operator-74547568cd-s95w5\" (UID: \"de5396bf-4d94-46be-87e1-7b587147a6a4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s95w5" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.660016 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ttn9\" (UniqueName: \"kubernetes.io/projected/de5396bf-4d94-46be-87e1-7b587147a6a4-kube-api-access-5ttn9\") pod \"machine-config-operator-74547568cd-s95w5\" (UID: \"de5396bf-4d94-46be-87e1-7b587147a6a4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s95w5" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.660049 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/24ccfcc4-25a6-4264-86b1-6ba0097b2614-trusted-ca\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.660067 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b748d317-5cb9-4fd3-a6bc-4a83d2f77d97-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lxbgb\" (UID: \"b748d317-5cb9-4fd3-a6bc-4a83d2f77d97\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lxbgb" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.660117 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dcc8e980-62b6-48fb-8339-fab83032bd0a-service-ca-bundle\") pod \"authentication-operator-69f744f599-82gdf\" (UID: \"dcc8e980-62b6-48fb-8339-fab83032bd0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-82gdf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.660223 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.660351 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/47a07011-2fae-40a2-9fb1-fec3461d2352-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-ddcbg\" (UID: \"47a07011-2fae-40a2-9fb1-fec3461d2352\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ddcbg" Oct 07 13:47:39 crc kubenswrapper[4959]: E1007 13:47:39.660556 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:40.16054484 +0000 UTC m=+142.243949165 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.660739 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47a07011-2fae-40a2-9fb1-fec3461d2352-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-ddcbg\" (UID: \"47a07011-2fae-40a2-9fb1-fec3461d2352\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ddcbg" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.660779 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dcc8e980-62b6-48fb-8339-fab83032bd0a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-82gdf\" (UID: \"dcc8e980-62b6-48fb-8339-fab83032bd0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-82gdf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.660816 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d9c7700-8ba2-444e-a3f0-5c5c6dd05585-config\") pod \"machine-api-operator-5694c8668f-dt2k4\" (UID: \"1d9c7700-8ba2-444e-a3f0-5c5c6dd05585\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dt2k4" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.660864 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a2a7f86-33ef-42d0-93b2-765b74d5c629-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-mx6p7\" (UID: \"4a2a7f86-33ef-42d0-93b2-765b74d5c629\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mx6p7" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.660886 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v28wj\" (UniqueName: \"kubernetes.io/projected/b816c481-6891-479c-99db-95f6e4abbc05-kube-api-access-v28wj\") pod \"package-server-manager-789f6589d5-94zbw\" (UID: \"b816c481-6891-479c-99db-95f6e4abbc05\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-94zbw" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.660906 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a2a7f86-33ef-42d0-93b2-765b74d5c629-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-mx6p7\" (UID: \"4a2a7f86-33ef-42d0-93b2-765b74d5c629\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mx6p7" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.660929 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33deb07b-91b4-45f1-88d4-deed38a97f36-config\") pod \"kube-apiserver-operator-766d6c64bb-zwd57\" (UID: \"33deb07b-91b4-45f1-88d4-deed38a97f36\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zwd57" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.661042 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e7167382-2267-4848-a423-e627253c38ed-node-pullsecrets\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.661063 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7167382-2267-4848-a423-e627253c38ed-config\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.661144 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b748d317-5cb9-4fd3-a6bc-4a83d2f77d97-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lxbgb\" (UID: \"b748d317-5cb9-4fd3-a6bc-4a83d2f77d97\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lxbgb" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.661168 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcc8e980-62b6-48fb-8339-fab83032bd0a-config\") pod \"authentication-operator-69f744f599-82gdf\" (UID: \"dcc8e980-62b6-48fb-8339-fab83032bd0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-82gdf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.661190 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dcc8e980-62b6-48fb-8339-fab83032bd0a-serving-cert\") pod \"authentication-operator-69f744f599-82gdf\" (UID: \"dcc8e980-62b6-48fb-8339-fab83032bd0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-82gdf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.661225 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/24ccfcc4-25a6-4264-86b1-6ba0097b2614-registry-certificates\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.661247 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f25b8da6-8185-4faa-b800-3deef3a9b06b-available-featuregates\") pod \"openshift-config-operator-7777fb866f-xkp55\" (UID: \"f25b8da6-8185-4faa-b800-3deef3a9b06b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xkp55" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.662008 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q75c2\" (UniqueName: \"kubernetes.io/projected/f25b8da6-8185-4faa-b800-3deef3a9b06b-kube-api-access-q75c2\") pod \"openshift-config-operator-7777fb866f-xkp55\" (UID: \"f25b8da6-8185-4faa-b800-3deef3a9b06b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xkp55" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.718359 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-qr2pg" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.756771 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8k48v" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.764817 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:39 crc kubenswrapper[4959]: E1007 13:47:39.764941 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:40.264920843 +0000 UTC m=+142.348325178 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765139 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f25b8da6-8185-4faa-b800-3deef3a9b06b-serving-cert\") pod \"openshift-config-operator-7777fb866f-xkp55\" (UID: \"f25b8da6-8185-4faa-b800-3deef3a9b06b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xkp55" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765180 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/bd1bdacf-99ea-4afc-b804-2a87afb35cc7-srv-cert\") pod \"olm-operator-6b444d44fb-ls9ng\" (UID: \"bd1bdacf-99ea-4afc-b804-2a87afb35cc7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ls9ng" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765202 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/62481657-1b38-4d8d-81d5-665bd2574467-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-t8v4d\" (UID: \"62481657-1b38-4d8d-81d5-665bd2574467\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t8v4d" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765228 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kljdw\" (UniqueName: \"kubernetes.io/projected/62481657-1b38-4d8d-81d5-665bd2574467-kube-api-access-kljdw\") pod \"cluster-image-registry-operator-dc59b4c8b-t8v4d\" (UID: \"62481657-1b38-4d8d-81d5-665bd2574467\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t8v4d" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765275 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b748d317-5cb9-4fd3-a6bc-4a83d2f77d97-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lxbgb\" (UID: \"b748d317-5cb9-4fd3-a6bc-4a83d2f77d97\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lxbgb" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765296 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmk25\" (UniqueName: \"kubernetes.io/projected/fe29debd-926a-4fd7-900c-b91a4d203436-kube-api-access-jmk25\") pod \"cluster-samples-operator-665b6dd947-2bh9c\" (UID: \"fe29debd-926a-4fd7-900c-b91a4d203436\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2bh9c" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765319 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47a07011-2fae-40a2-9fb1-fec3461d2352-config\") pod \"kube-controller-manager-operator-78b949d7b-ddcbg\" (UID: \"47a07011-2fae-40a2-9fb1-fec3461d2352\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ddcbg" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765342 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/24ccfcc4-25a6-4264-86b1-6ba0097b2614-installation-pull-secrets\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765361 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cl266\" (UniqueName: \"kubernetes.io/projected/9872926d-9b2f-489a-804d-749ec1e2c136-kube-api-access-cl266\") pod \"service-ca-operator-777779d784-q9cfv\" (UID: \"9872926d-9b2f-489a-804d-749ec1e2c136\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-q9cfv" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765385 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs2fh\" (UniqueName: \"kubernetes.io/projected/e7167382-2267-4848-a423-e627253c38ed-kube-api-access-rs2fh\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765419 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/24ccfcc4-25a6-4264-86b1-6ba0097b2614-bound-sa-token\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765441 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3218d84a-b760-4002-bf1f-79d02293ad93-metrics-tls\") pod \"dns-operator-744455d44c-cw2cl\" (UID: \"3218d84a-b760-4002-bf1f-79d02293ad93\") " pod="openshift-dns-operator/dns-operator-744455d44c-cw2cl" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765462 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/bd1bdacf-99ea-4afc-b804-2a87afb35cc7-profile-collector-cert\") pod \"olm-operator-6b444d44fb-ls9ng\" (UID: \"bd1bdacf-99ea-4afc-b804-2a87afb35cc7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ls9ng" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765530 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81d176a3-10ba-42f5-b716-d18a7dc1b5a8-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zg2sm\" (UID: \"81d176a3-10ba-42f5-b716-d18a7dc1b5a8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zg2sm" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765554 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2843a931-4390-4dae-a736-911e2fc18bef-srv-cert\") pod \"catalog-operator-68c6474976-gfnhv\" (UID: \"2843a931-4390-4dae-a736-911e2fc18bef\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gfnhv" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765582 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/24ccfcc4-25a6-4264-86b1-6ba0097b2614-ca-trust-extracted\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765605 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/de5396bf-4d94-46be-87e1-7b587147a6a4-proxy-tls\") pod \"machine-config-operator-74547568cd-s95w5\" (UID: \"de5396bf-4d94-46be-87e1-7b587147a6a4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s95w5" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765640 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ttn9\" (UniqueName: \"kubernetes.io/projected/de5396bf-4d94-46be-87e1-7b587147a6a4-kube-api-access-5ttn9\") pod \"machine-config-operator-74547568cd-s95w5\" (UID: \"de5396bf-4d94-46be-87e1-7b587147a6a4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s95w5" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765661 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7167382-2267-4848-a423-e627253c38ed-serving-cert\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765696 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b748d317-5cb9-4fd3-a6bc-4a83d2f77d97-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lxbgb\" (UID: \"b748d317-5cb9-4fd3-a6bc-4a83d2f77d97\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lxbgb" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765719 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dcc8e980-62b6-48fb-8339-fab83032bd0a-service-ca-bundle\") pod \"authentication-operator-69f744f599-82gdf\" (UID: \"dcc8e980-62b6-48fb-8339-fab83032bd0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-82gdf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765753 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e7167382-2267-4848-a423-e627253c38ed-etcd-serving-ca\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765777 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/3b0dd795-8b21-43a2-9e50-b0cb668ac22c-tmpfs\") pod \"packageserver-d55dfcdfc-5z54z\" (UID: \"3b0dd795-8b21-43a2-9e50-b0cb668ac22c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z54z" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765804 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765827 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/47a07011-2fae-40a2-9fb1-fec3461d2352-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-ddcbg\" (UID: \"47a07011-2fae-40a2-9fb1-fec3461d2352\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ddcbg" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765849 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47a07011-2fae-40a2-9fb1-fec3461d2352-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-ddcbg\" (UID: \"47a07011-2fae-40a2-9fb1-fec3461d2352\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ddcbg" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765885 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7526c\" (UniqueName: \"kubernetes.io/projected/3218d84a-b760-4002-bf1f-79d02293ad93-kube-api-access-7526c\") pod \"dns-operator-744455d44c-cw2cl\" (UID: \"3218d84a-b760-4002-bf1f-79d02293ad93\") " pod="openshift-dns-operator/dns-operator-744455d44c-cw2cl" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765921 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cb8de785-a298-414c-998c-7bd0585966e1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xwlzm\" (UID: \"cb8de785-a298-414c-998c-7bd0585966e1\") " pod="openshift-marketplace/marketplace-operator-79b997595-xwlzm" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765944 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d9c7700-8ba2-444e-a3f0-5c5c6dd05585-config\") pod \"machine-api-operator-5694c8668f-dt2k4\" (UID: \"1d9c7700-8ba2-444e-a3f0-5c5c6dd05585\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dt2k4" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765966 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a2a7f86-33ef-42d0-93b2-765b74d5c629-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-mx6p7\" (UID: \"4a2a7f86-33ef-42d0-93b2-765b74d5c629\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mx6p7" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.765988 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/468622a9-56e9-47f6-aa19-b9df629881b7-signing-cabundle\") pod \"service-ca-9c57cc56f-v5pq6\" (UID: \"468622a9-56e9-47f6-aa19-b9df629881b7\") " pod="openshift-service-ca/service-ca-9c57cc56f-v5pq6" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766010 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a2a7f86-33ef-42d0-93b2-765b74d5c629-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-mx6p7\" (UID: \"4a2a7f86-33ef-42d0-93b2-765b74d5c629\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mx6p7" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766031 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e7167382-2267-4848-a423-e627253c38ed-audit-dir\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766053 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e7167382-2267-4848-a423-e627253c38ed-node-pullsecrets\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766090 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b748d317-5cb9-4fd3-a6bc-4a83d2f77d97-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lxbgb\" (UID: \"b748d317-5cb9-4fd3-a6bc-4a83d2f77d97\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lxbgb" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766130 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dcc8e980-62b6-48fb-8339-fab83032bd0a-serving-cert\") pod \"authentication-operator-69f744f599-82gdf\" (UID: \"dcc8e980-62b6-48fb-8339-fab83032bd0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-82gdf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766155 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/62481657-1b38-4d8d-81d5-665bd2574467-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-t8v4d\" (UID: \"62481657-1b38-4d8d-81d5-665bd2574467\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t8v4d" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766199 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/fd9598df-e972-4ee1-bfa4-e509f6bd04bc-certs\") pod \"machine-config-server-xn2j9\" (UID: \"fd9598df-e972-4ee1-bfa4-e509f6bd04bc\") " pod="openshift-machine-config-operator/machine-config-server-xn2j9" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766235 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrbcf\" (UniqueName: \"kubernetes.io/projected/dcc8e980-62b6-48fb-8339-fab83032bd0a-kube-api-access-qrbcf\") pod \"authentication-operator-69f744f599-82gdf\" (UID: \"dcc8e980-62b6-48fb-8339-fab83032bd0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-82gdf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766292 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/fe29debd-926a-4fd7-900c-b91a4d203436-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-2bh9c\" (UID: \"fe29debd-926a-4fd7-900c-b91a4d203436\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2bh9c" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766314 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c286e288-6157-4fff-bb4a-927e823d451a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-jj9rd\" (UID: \"c286e288-6157-4fff-bb4a-927e823d451a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jj9rd" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766336 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/62481657-1b38-4d8d-81d5-665bd2574467-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-t8v4d\" (UID: \"62481657-1b38-4d8d-81d5-665bd2574467\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t8v4d" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766361 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsm8h\" (UniqueName: \"kubernetes.io/projected/24ccfcc4-25a6-4264-86b1-6ba0097b2614-kube-api-access-xsm8h\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766383 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c286e288-6157-4fff-bb4a-927e823d451a-metrics-tls\") pod \"ingress-operator-5b745b69d9-jj9rd\" (UID: \"c286e288-6157-4fff-bb4a-927e823d451a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jj9rd" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766404 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e7167382-2267-4848-a423-e627253c38ed-image-import-ca\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766444 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9872926d-9b2f-489a-804d-749ec1e2c136-serving-cert\") pod \"service-ca-operator-777779d784-q9cfv\" (UID: \"9872926d-9b2f-489a-804d-749ec1e2c136\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-q9cfv" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766467 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k25t2\" (UniqueName: \"kubernetes.io/projected/4a2a7f86-33ef-42d0-93b2-765b74d5c629-kube-api-access-k25t2\") pod \"openshift-controller-manager-operator-756b6f6bc6-mx6p7\" (UID: \"4a2a7f86-33ef-42d0-93b2-765b74d5c629\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mx6p7" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766490 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3b0dd795-8b21-43a2-9e50-b0cb668ac22c-webhook-cert\") pod \"packageserver-d55dfcdfc-5z54z\" (UID: \"3b0dd795-8b21-43a2-9e50-b0cb668ac22c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z54z" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766511 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/09af6fe0-f3ab-4223-9b45-05509e575ac4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-5qkt6\" (UID: \"09af6fe0-f3ab-4223-9b45-05509e575ac4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5qkt6" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766535 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54540da0-9e56-4ce2-bf77-1d055296e4b1-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rxt22\" (UID: \"54540da0-9e56-4ce2-bf77-1d055296e4b1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rxt22" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766667 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0a596b0-4162-4b10-8880-3ea0421b8c74-secret-volume\") pod \"collect-profiles-29330745-wbwkx\" (UID: \"d0a596b0-4162-4b10-8880-3ea0421b8c74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330745-wbwkx" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766691 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/fd9598df-e972-4ee1-bfa4-e509f6bd04bc-node-bootstrap-token\") pod \"machine-config-server-xn2j9\" (UID: \"fd9598df-e972-4ee1-bfa4-e509f6bd04bc\") " pod="openshift-machine-config-operator/machine-config-server-xn2j9" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766716 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b816c481-6891-479c-99db-95f6e4abbc05-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-94zbw\" (UID: \"b816c481-6891-479c-99db-95f6e4abbc05\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-94zbw" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766737 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1d9c7700-8ba2-444e-a3f0-5c5c6dd05585-images\") pod \"machine-api-operator-5694c8668f-dt2k4\" (UID: \"1d9c7700-8ba2-444e-a3f0-5c5c6dd05585\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dt2k4" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766761 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6r8tg\" (UniqueName: \"kubernetes.io/projected/09af6fe0-f3ab-4223-9b45-05509e575ac4-kube-api-access-6r8tg\") pod \"multus-admission-controller-857f4d67dd-5qkt6\" (UID: \"09af6fe0-f3ab-4223-9b45-05509e575ac4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5qkt6" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766808 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/33deb07b-91b4-45f1-88d4-deed38a97f36-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-zwd57\" (UID: \"33deb07b-91b4-45f1-88d4-deed38a97f36\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zwd57" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766847 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9d4h\" (UniqueName: \"kubernetes.io/projected/d0a596b0-4162-4b10-8880-3ea0421b8c74-kube-api-access-j9d4h\") pod \"collect-profiles-29330745-wbwkx\" (UID: \"d0a596b0-4162-4b10-8880-3ea0421b8c74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330745-wbwkx" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766895 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/33deb07b-91b4-45f1-88d4-deed38a97f36-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-zwd57\" (UID: \"33deb07b-91b4-45f1-88d4-deed38a97f36\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zwd57" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766919 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/1d9c7700-8ba2-444e-a3f0-5c5c6dd05585-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-dt2k4\" (UID: \"1d9c7700-8ba2-444e-a3f0-5c5c6dd05585\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dt2k4" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766942 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9dwg\" (UniqueName: \"kubernetes.io/projected/1d9c7700-8ba2-444e-a3f0-5c5c6dd05585-kube-api-access-d9dwg\") pod \"machine-api-operator-5694c8668f-dt2k4\" (UID: \"1d9c7700-8ba2-444e-a3f0-5c5c6dd05585\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dt2k4" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.766964 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e7167382-2267-4848-a423-e627253c38ed-trusted-ca-bundle\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.767000 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hqx5\" (UniqueName: \"kubernetes.io/projected/81d176a3-10ba-42f5-b716-d18a7dc1b5a8-kube-api-access-5hqx5\") pod \"openshift-apiserver-operator-796bbdcf4f-zg2sm\" (UID: \"81d176a3-10ba-42f5-b716-d18a7dc1b5a8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zg2sm" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.767026 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9872926d-9b2f-489a-804d-749ec1e2c136-config\") pod \"service-ca-operator-777779d784-q9cfv\" (UID: \"9872926d-9b2f-489a-804d-749ec1e2c136\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-q9cfv" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.767049 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d6vt\" (UniqueName: \"kubernetes.io/projected/cb8de785-a298-414c-998c-7bd0585966e1-kube-api-access-6d6vt\") pod \"marketplace-operator-79b997595-xwlzm\" (UID: \"cb8de785-a298-414c-998c-7bd0585966e1\") " pod="openshift-marketplace/marketplace-operator-79b997595-xwlzm" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.767072 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/24ccfcc4-25a6-4264-86b1-6ba0097b2614-registry-tls\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.767094 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/de5396bf-4d94-46be-87e1-7b587147a6a4-images\") pod \"machine-config-operator-74547568cd-s95w5\" (UID: \"de5396bf-4d94-46be-87e1-7b587147a6a4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s95w5" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.767384 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81d176a3-10ba-42f5-b716-d18a7dc1b5a8-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zg2sm\" (UID: \"81d176a3-10ba-42f5-b716-d18a7dc1b5a8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zg2sm" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.768085 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/de5396bf-4d94-46be-87e1-7b587147a6a4-images\") pod \"machine-config-operator-74547568cd-s95w5\" (UID: \"de5396bf-4d94-46be-87e1-7b587147a6a4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s95w5" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.768172 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e7167382-2267-4848-a423-e627253c38ed-node-pullsecrets\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.768750 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a2a7f86-33ef-42d0-93b2-765b74d5c629-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-mx6p7\" (UID: \"4a2a7f86-33ef-42d0-93b2-765b74d5c629\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mx6p7" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.768796 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/15866f74-4fd6-46df-b182-7cd58b5c1e0b-cert\") pod \"ingress-canary-7cvhs\" (UID: \"15866f74-4fd6-46df-b182-7cd58b5c1e0b\") " pod="openshift-ingress-canary/ingress-canary-7cvhs" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.768844 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e7167382-2267-4848-a423-e627253c38ed-encryption-config\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.768909 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/24ccfcc4-25a6-4264-86b1-6ba0097b2614-trusted-ca\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.768957 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54540da0-9e56-4ce2-bf77-1d055296e4b1-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rxt22\" (UID: \"54540da0-9e56-4ce2-bf77-1d055296e4b1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rxt22" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.768982 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/468622a9-56e9-47f6-aa19-b9df629881b7-signing-key\") pod \"service-ca-9c57cc56f-v5pq6\" (UID: \"468622a9-56e9-47f6-aa19-b9df629881b7\") " pod="openshift-service-ca/service-ca-9c57cc56f-v5pq6" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.769067 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pk9fk\" (UniqueName: \"kubernetes.io/projected/932c7fc9-8796-41bc-982f-74dfd0d887be-kube-api-access-pk9fk\") pod \"machine-config-controller-84d6567774-529ml\" (UID: \"932c7fc9-8796-41bc-982f-74dfd0d887be\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-529ml" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.769142 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dcc8e980-62b6-48fb-8339-fab83032bd0a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-82gdf\" (UID: \"dcc8e980-62b6-48fb-8339-fab83032bd0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-82gdf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.769187 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v28wj\" (UniqueName: \"kubernetes.io/projected/b816c481-6891-479c-99db-95f6e4abbc05-kube-api-access-v28wj\") pod \"package-server-manager-789f6589d5-94zbw\" (UID: \"b816c481-6891-479c-99db-95f6e4abbc05\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-94zbw" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.769218 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/932c7fc9-8796-41bc-982f-74dfd0d887be-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-529ml\" (UID: \"932c7fc9-8796-41bc-982f-74dfd0d887be\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-529ml" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.769248 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33deb07b-91b4-45f1-88d4-deed38a97f36-config\") pod \"kube-apiserver-operator-766d6c64bb-zwd57\" (UID: \"33deb07b-91b4-45f1-88d4-deed38a97f36\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zwd57" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.771055 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47a07011-2fae-40a2-9fb1-fec3461d2352-config\") pod \"kube-controller-manager-operator-78b949d7b-ddcbg\" (UID: \"47a07011-2fae-40a2-9fb1-fec3461d2352\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ddcbg" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.771120 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-phxkp" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.771239 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1d9c7700-8ba2-444e-a3f0-5c5c6dd05585-images\") pod \"machine-api-operator-5694c8668f-dt2k4\" (UID: \"1d9c7700-8ba2-444e-a3f0-5c5c6dd05585\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dt2k4" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.771275 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7167382-2267-4848-a423-e627253c38ed-config\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.771319 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2qlj\" (UniqueName: \"kubernetes.io/projected/54540da0-9e56-4ce2-bf77-1d055296e4b1-kube-api-access-v2qlj\") pod \"kube-storage-version-migrator-operator-b67b599dd-rxt22\" (UID: \"54540da0-9e56-4ce2-bf77-1d055296e4b1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rxt22" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.771362 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9kbr\" (UniqueName: \"kubernetes.io/projected/fd9598df-e972-4ee1-bfa4-e509f6bd04bc-kube-api-access-b9kbr\") pod \"machine-config-server-xn2j9\" (UID: \"fd9598df-e972-4ee1-bfa4-e509f6bd04bc\") " pod="openshift-machine-config-operator/machine-config-server-xn2j9" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.771421 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djztn\" (UniqueName: \"kubernetes.io/projected/bd1bdacf-99ea-4afc-b804-2a87afb35cc7-kube-api-access-djztn\") pod \"olm-operator-6b444d44fb-ls9ng\" (UID: \"bd1bdacf-99ea-4afc-b804-2a87afb35cc7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ls9ng" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.771447 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zt5s\" (UniqueName: \"kubernetes.io/projected/468622a9-56e9-47f6-aa19-b9df629881b7-kube-api-access-2zt5s\") pod \"service-ca-9c57cc56f-v5pq6\" (UID: \"468622a9-56e9-47f6-aa19-b9df629881b7\") " pod="openshift-service-ca/service-ca-9c57cc56f-v5pq6" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.771473 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0a596b0-4162-4b10-8880-3ea0421b8c74-config-volume\") pod \"collect-profiles-29330745-wbwkx\" (UID: \"d0a596b0-4162-4b10-8880-3ea0421b8c74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330745-wbwkx" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.771533 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcc8e980-62b6-48fb-8339-fab83032bd0a-config\") pod \"authentication-operator-69f744f599-82gdf\" (UID: \"dcc8e980-62b6-48fb-8339-fab83032bd0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-82gdf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.771571 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f25b8da6-8185-4faa-b800-3deef3a9b06b-available-featuregates\") pod \"openshift-config-operator-7777fb866f-xkp55\" (UID: \"f25b8da6-8185-4faa-b800-3deef3a9b06b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xkp55" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.774792 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7167382-2267-4848-a423-e627253c38ed-config\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.775154 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/33deb07b-91b4-45f1-88d4-deed38a97f36-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-zwd57\" (UID: \"33deb07b-91b4-45f1-88d4-deed38a97f36\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zwd57" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.775607 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b748d317-5cb9-4fd3-a6bc-4a83d2f77d97-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lxbgb\" (UID: \"b748d317-5cb9-4fd3-a6bc-4a83d2f77d97\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lxbgb" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.775690 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/24ccfcc4-25a6-4264-86b1-6ba0097b2614-ca-trust-extracted\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.775921 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/24ccfcc4-25a6-4264-86b1-6ba0097b2614-installation-pull-secrets\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.776590 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxl5f\" (UniqueName: \"kubernetes.io/projected/8307e459-035b-4c33-bb8e-b5a698618036-kube-api-access-xxl5f\") pod \"downloads-7954f5f757-7bxtq\" (UID: \"8307e459-035b-4c33-bb8e-b5a698618036\") " pod="openshift-console/downloads-7954f5f757-7bxtq" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.776640 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/24ccfcc4-25a6-4264-86b1-6ba0097b2614-registry-certificates\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.776676 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e7167382-2267-4848-a423-e627253c38ed-audit\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.776706 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnhf6\" (UniqueName: \"kubernetes.io/projected/15866f74-4fd6-46df-b182-7cd58b5c1e0b-kube-api-access-dnhf6\") pod \"ingress-canary-7cvhs\" (UID: \"15866f74-4fd6-46df-b182-7cd58b5c1e0b\") " pod="openshift-ingress-canary/ingress-canary-7cvhs" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.776858 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q75c2\" (UniqueName: \"kubernetes.io/projected/f25b8da6-8185-4faa-b800-3deef3a9b06b-kube-api-access-q75c2\") pod \"openshift-config-operator-7777fb866f-xkp55\" (UID: \"f25b8da6-8185-4faa-b800-3deef3a9b06b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xkp55" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.777001 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3b0dd795-8b21-43a2-9e50-b0cb668ac22c-apiservice-cert\") pod \"packageserver-d55dfcdfc-5z54z\" (UID: \"3b0dd795-8b21-43a2-9e50-b0cb668ac22c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z54z" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.777051 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/cb8de785-a298-414c-998c-7bd0585966e1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xwlzm\" (UID: \"cb8de785-a298-414c-998c-7bd0585966e1\") " pod="openshift-marketplace/marketplace-operator-79b997595-xwlzm" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.777245 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e7167382-2267-4848-a423-e627253c38ed-etcd-client\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.777284 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/de5396bf-4d94-46be-87e1-7b587147a6a4-auth-proxy-config\") pod \"machine-config-operator-74547568cd-s95w5\" (UID: \"de5396bf-4d94-46be-87e1-7b587147a6a4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s95w5" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.777319 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c286e288-6157-4fff-bb4a-927e823d451a-trusted-ca\") pod \"ingress-operator-5b745b69d9-jj9rd\" (UID: \"c286e288-6157-4fff-bb4a-927e823d451a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jj9rd" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.778772 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdxq7\" (UniqueName: \"kubernetes.io/projected/c286e288-6157-4fff-bb4a-927e823d451a-kube-api-access-xdxq7\") pod \"ingress-operator-5b745b69d9-jj9rd\" (UID: \"c286e288-6157-4fff-bb4a-927e823d451a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jj9rd" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.778858 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/932c7fc9-8796-41bc-982f-74dfd0d887be-proxy-tls\") pod \"machine-config-controller-84d6567774-529ml\" (UID: \"932c7fc9-8796-41bc-982f-74dfd0d887be\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-529ml" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.778973 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5m9gl\" (UniqueName: \"kubernetes.io/projected/3b0dd795-8b21-43a2-9e50-b0cb668ac22c-kube-api-access-5m9gl\") pod \"packageserver-d55dfcdfc-5z54z\" (UID: \"3b0dd795-8b21-43a2-9e50-b0cb668ac22c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z54z" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.779942 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33deb07b-91b4-45f1-88d4-deed38a97f36-config\") pod \"kube-apiserver-operator-766d6c64bb-zwd57\" (UID: \"33deb07b-91b4-45f1-88d4-deed38a97f36\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zwd57" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.776723 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9872926d-9b2f-489a-804d-749ec1e2c136-serving-cert\") pod \"service-ca-operator-777779d784-q9cfv\" (UID: \"9872926d-9b2f-489a-804d-749ec1e2c136\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-q9cfv" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.780726 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9872926d-9b2f-489a-804d-749ec1e2c136-config\") pod \"service-ca-operator-777779d784-q9cfv\" (UID: \"9872926d-9b2f-489a-804d-749ec1e2c136\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-q9cfv" Oct 07 13:47:39 crc kubenswrapper[4959]: E1007 13:47:39.781791 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:40.281767138 +0000 UTC m=+142.365171463 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.782010 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/de5396bf-4d94-46be-87e1-7b587147a6a4-proxy-tls\") pod \"machine-config-operator-74547568cd-s95w5\" (UID: \"de5396bf-4d94-46be-87e1-7b587147a6a4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s95w5" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.782656 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b748d317-5cb9-4fd3-a6bc-4a83d2f77d97-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lxbgb\" (UID: \"b748d317-5cb9-4fd3-a6bc-4a83d2f77d97\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lxbgb" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.783264 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dcc8e980-62b6-48fb-8339-fab83032bd0a-service-ca-bundle\") pod \"authentication-operator-69f744f599-82gdf\" (UID: \"dcc8e980-62b6-48fb-8339-fab83032bd0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-82gdf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.785367 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f25b8da6-8185-4faa-b800-3deef3a9b06b-available-featuregates\") pod \"openshift-config-operator-7777fb866f-xkp55\" (UID: \"f25b8da6-8185-4faa-b800-3deef3a9b06b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xkp55" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.785481 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcc8e980-62b6-48fb-8339-fab83032bd0a-config\") pod \"authentication-operator-69f744f599-82gdf\" (UID: \"dcc8e980-62b6-48fb-8339-fab83032bd0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-82gdf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.786382 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/de5396bf-4d94-46be-87e1-7b587147a6a4-auth-proxy-config\") pod \"machine-config-operator-74547568cd-s95w5\" (UID: \"de5396bf-4d94-46be-87e1-7b587147a6a4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s95w5" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.788278 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/24ccfcc4-25a6-4264-86b1-6ba0097b2614-registry-certificates\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.789066 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d9c7700-8ba2-444e-a3f0-5c5c6dd05585-config\") pod \"machine-api-operator-5694c8668f-dt2k4\" (UID: \"1d9c7700-8ba2-444e-a3f0-5c5c6dd05585\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dt2k4" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.789647 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/24ccfcc4-25a6-4264-86b1-6ba0097b2614-trusted-ca\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.789873 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dcc8e980-62b6-48fb-8339-fab83032bd0a-serving-cert\") pod \"authentication-operator-69f744f599-82gdf\" (UID: \"dcc8e980-62b6-48fb-8339-fab83032bd0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-82gdf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.790084 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/1d9c7700-8ba2-444e-a3f0-5c5c6dd05585-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-dt2k4\" (UID: \"1d9c7700-8ba2-444e-a3f0-5c5c6dd05585\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dt2k4" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.793167 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47a07011-2fae-40a2-9fb1-fec3461d2352-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-ddcbg\" (UID: \"47a07011-2fae-40a2-9fb1-fec3461d2352\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ddcbg" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.797635 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-pcvlp" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.797667 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/24ccfcc4-25a6-4264-86b1-6ba0097b2614-registry-tls\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.798487 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f25b8da6-8185-4faa-b800-3deef3a9b06b-serving-cert\") pod \"openshift-config-operator-7777fb866f-xkp55\" (UID: \"f25b8da6-8185-4faa-b800-3deef3a9b06b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xkp55" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.799250 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dcc8e980-62b6-48fb-8339-fab83032bd0a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-82gdf\" (UID: \"dcc8e980-62b6-48fb-8339-fab83032bd0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-82gdf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.802176 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b816c481-6891-479c-99db-95f6e4abbc05-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-94zbw\" (UID: \"b816c481-6891-479c-99db-95f6e4abbc05\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-94zbw" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.805927 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a2a7f86-33ef-42d0-93b2-765b74d5c629-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-mx6p7\" (UID: \"4a2a7f86-33ef-42d0-93b2-765b74d5c629\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mx6p7" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.806173 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-8nt7g" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.806233 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/fe29debd-926a-4fd7-900c-b91a4d203436-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-2bh9c\" (UID: \"fe29debd-926a-4fd7-900c-b91a4d203436\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2bh9c" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.825010 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/24ccfcc4-25a6-4264-86b1-6ba0097b2614-bound-sa-token\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.829691 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cl266\" (UniqueName: \"kubernetes.io/projected/9872926d-9b2f-489a-804d-749ec1e2c136-kube-api-access-cl266\") pod \"service-ca-operator-777779d784-q9cfv\" (UID: \"9872926d-9b2f-489a-804d-749ec1e2c136\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-q9cfv" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.843671 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jdvrw" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.850064 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k25t2\" (UniqueName: \"kubernetes.io/projected/4a2a7f86-33ef-42d0-93b2-765b74d5c629-kube-api-access-k25t2\") pod \"openshift-controller-manager-operator-756b6f6bc6-mx6p7\" (UID: \"4a2a7f86-33ef-42d0-93b2-765b74d5c629\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mx6p7" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.864142 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-q9cfv" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.871977 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsm8h\" (UniqueName: \"kubernetes.io/projected/24ccfcc4-25a6-4264-86b1-6ba0097b2614-kube-api-access-xsm8h\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.888973 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.889247 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/07357227-cc26-43b8-89c7-4990980b2725-registration-dir\") pod \"csi-hostpathplugin-fggsp\" (UID: \"07357227-cc26-43b8-89c7-4990980b2725\") " pod="hostpath-provisioner/csi-hostpathplugin-fggsp" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.889290 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs2fh\" (UniqueName: \"kubernetes.io/projected/e7167382-2267-4848-a423-e627253c38ed-kube-api-access-rs2fh\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.889445 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3218d84a-b760-4002-bf1f-79d02293ad93-metrics-tls\") pod \"dns-operator-744455d44c-cw2cl\" (UID: \"3218d84a-b760-4002-bf1f-79d02293ad93\") " pod="openshift-dns-operator/dns-operator-744455d44c-cw2cl" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.889482 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/bd1bdacf-99ea-4afc-b804-2a87afb35cc7-profile-collector-cert\") pod \"olm-operator-6b444d44fb-ls9ng\" (UID: \"bd1bdacf-99ea-4afc-b804-2a87afb35cc7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ls9ng" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.889502 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81d176a3-10ba-42f5-b716-d18a7dc1b5a8-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zg2sm\" (UID: \"81d176a3-10ba-42f5-b716-d18a7dc1b5a8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zg2sm" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.889527 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/07357227-cc26-43b8-89c7-4990980b2725-plugins-dir\") pod \"csi-hostpathplugin-fggsp\" (UID: \"07357227-cc26-43b8-89c7-4990980b2725\") " pod="hostpath-provisioner/csi-hostpathplugin-fggsp" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.889552 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2843a931-4390-4dae-a736-911e2fc18bef-srv-cert\") pod \"catalog-operator-68c6474976-gfnhv\" (UID: \"2843a931-4390-4dae-a736-911e2fc18bef\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gfnhv" Oct 07 13:47:39 crc kubenswrapper[4959]: E1007 13:47:39.890579 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:40.390555313 +0000 UTC m=+142.473959638 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.891010 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ttn9\" (UniqueName: \"kubernetes.io/projected/de5396bf-4d94-46be-87e1-7b587147a6a4-kube-api-access-5ttn9\") pod \"machine-config-operator-74547568cd-s95w5\" (UID: \"de5396bf-4d94-46be-87e1-7b587147a6a4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s95w5" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.891341 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2843a931-4390-4dae-a736-911e2fc18bef-profile-collector-cert\") pod \"catalog-operator-68c6474976-gfnhv\" (UID: \"2843a931-4390-4dae-a736-911e2fc18bef\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gfnhv" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.891378 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7167382-2267-4848-a423-e627253c38ed-serving-cert\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.891433 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.891469 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e7167382-2267-4848-a423-e627253c38ed-etcd-serving-ca\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.891494 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/3b0dd795-8b21-43a2-9e50-b0cb668ac22c-tmpfs\") pod \"packageserver-d55dfcdfc-5z54z\" (UID: \"3b0dd795-8b21-43a2-9e50-b0cb668ac22c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z54z" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.891521 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7526c\" (UniqueName: \"kubernetes.io/projected/3218d84a-b760-4002-bf1f-79d02293ad93-kube-api-access-7526c\") pod \"dns-operator-744455d44c-cw2cl\" (UID: \"3218d84a-b760-4002-bf1f-79d02293ad93\") " pod="openshift-dns-operator/dns-operator-744455d44c-cw2cl" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.891546 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cb8de785-a298-414c-998c-7bd0585966e1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xwlzm\" (UID: \"cb8de785-a298-414c-998c-7bd0585966e1\") " pod="openshift-marketplace/marketplace-operator-79b997595-xwlzm" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.891572 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/468622a9-56e9-47f6-aa19-b9df629881b7-signing-cabundle\") pod \"service-ca-9c57cc56f-v5pq6\" (UID: \"468622a9-56e9-47f6-aa19-b9df629881b7\") " pod="openshift-service-ca/service-ca-9c57cc56f-v5pq6" Oct 07 13:47:39 crc kubenswrapper[4959]: E1007 13:47:39.891956 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:40.391933431 +0000 UTC m=+142.475337786 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.892421 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e7167382-2267-4848-a423-e627253c38ed-etcd-serving-ca\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.892684 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/3b0dd795-8b21-43a2-9e50-b0cb668ac22c-tmpfs\") pod \"packageserver-d55dfcdfc-5z54z\" (UID: \"3b0dd795-8b21-43a2-9e50-b0cb668ac22c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z54z" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.893155 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cb8de785-a298-414c-998c-7bd0585966e1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xwlzm\" (UID: \"cb8de785-a298-414c-998c-7bd0585966e1\") " pod="openshift-marketplace/marketplace-operator-79b997595-xwlzm" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.893894 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/468622a9-56e9-47f6-aa19-b9df629881b7-signing-cabundle\") pod \"service-ca-9c57cc56f-v5pq6\" (UID: \"468622a9-56e9-47f6-aa19-b9df629881b7\") " pod="openshift-service-ca/service-ca-9c57cc56f-v5pq6" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.896288 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e7167382-2267-4848-a423-e627253c38ed-audit-dir\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.896331 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/07357227-cc26-43b8-89c7-4990980b2725-mountpoint-dir\") pod \"csi-hostpathplugin-fggsp\" (UID: \"07357227-cc26-43b8-89c7-4990980b2725\") " pod="hostpath-provisioner/csi-hostpathplugin-fggsp" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.896376 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/62481657-1b38-4d8d-81d5-665bd2574467-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-t8v4d\" (UID: \"62481657-1b38-4d8d-81d5-665bd2574467\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t8v4d" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.896403 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/fd9598df-e972-4ee1-bfa4-e509f6bd04bc-certs\") pod \"machine-config-server-xn2j9\" (UID: \"fd9598df-e972-4ee1-bfa4-e509f6bd04bc\") " pod="openshift-machine-config-operator/machine-config-server-xn2j9" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.896426 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/48528215-02d5-48aa-8289-9defecf47166-metrics-tls\") pod \"dns-default-rfcqs\" (UID: \"48528215-02d5-48aa-8289-9defecf47166\") " pod="openshift-dns/dns-default-rfcqs" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.896458 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c286e288-6157-4fff-bb4a-927e823d451a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-jj9rd\" (UID: \"c286e288-6157-4fff-bb4a-927e823d451a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jj9rd" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.896488 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/62481657-1b38-4d8d-81d5-665bd2574467-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-t8v4d\" (UID: \"62481657-1b38-4d8d-81d5-665bd2574467\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t8v4d" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.896515 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c286e288-6157-4fff-bb4a-927e823d451a-metrics-tls\") pod \"ingress-operator-5b745b69d9-jj9rd\" (UID: \"c286e288-6157-4fff-bb4a-927e823d451a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jj9rd" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.896539 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e7167382-2267-4848-a423-e627253c38ed-image-import-ca\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.896563 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8qs8\" (UniqueName: \"kubernetes.io/projected/2843a931-4390-4dae-a736-911e2fc18bef-kube-api-access-k8qs8\") pod \"catalog-operator-68c6474976-gfnhv\" (UID: \"2843a931-4390-4dae-a736-911e2fc18bef\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gfnhv" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.896643 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81d176a3-10ba-42f5-b716-d18a7dc1b5a8-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zg2sm\" (UID: \"81d176a3-10ba-42f5-b716-d18a7dc1b5a8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zg2sm" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.897881 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/07357227-cc26-43b8-89c7-4990980b2725-csi-data-dir\") pod \"csi-hostpathplugin-fggsp\" (UID: \"07357227-cc26-43b8-89c7-4990980b2725\") " pod="hostpath-provisioner/csi-hostpathplugin-fggsp" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.897924 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/09af6fe0-f3ab-4223-9b45-05509e575ac4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-5qkt6\" (UID: \"09af6fe0-f3ab-4223-9b45-05509e575ac4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5qkt6" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.897986 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3b0dd795-8b21-43a2-9e50-b0cb668ac22c-webhook-cert\") pod \"packageserver-d55dfcdfc-5z54z\" (UID: \"3b0dd795-8b21-43a2-9e50-b0cb668ac22c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z54z" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.898612 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e7167382-2267-4848-a423-e627253c38ed-audit-dir\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.898766 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54540da0-9e56-4ce2-bf77-1d055296e4b1-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rxt22\" (UID: \"54540da0-9e56-4ce2-bf77-1d055296e4b1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rxt22" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.898802 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0a596b0-4162-4b10-8880-3ea0421b8c74-secret-volume\") pod \"collect-profiles-29330745-wbwkx\" (UID: \"d0a596b0-4162-4b10-8880-3ea0421b8c74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330745-wbwkx" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.898824 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/fd9598df-e972-4ee1-bfa4-e509f6bd04bc-node-bootstrap-token\") pod \"machine-config-server-xn2j9\" (UID: \"fd9598df-e972-4ee1-bfa4-e509f6bd04bc\") " pod="openshift-machine-config-operator/machine-config-server-xn2j9" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.898854 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7mcx\" (UniqueName: \"kubernetes.io/projected/48528215-02d5-48aa-8289-9defecf47166-kube-api-access-t7mcx\") pod \"dns-default-rfcqs\" (UID: \"48528215-02d5-48aa-8289-9defecf47166\") " pod="openshift-dns/dns-default-rfcqs" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.898888 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7167382-2267-4848-a423-e627253c38ed-serving-cert\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.898910 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6r8tg\" (UniqueName: \"kubernetes.io/projected/09af6fe0-f3ab-4223-9b45-05509e575ac4-kube-api-access-6r8tg\") pod \"multus-admission-controller-857f4d67dd-5qkt6\" (UID: \"09af6fe0-f3ab-4223-9b45-05509e575ac4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5qkt6" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.898980 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9d4h\" (UniqueName: \"kubernetes.io/projected/d0a596b0-4162-4b10-8880-3ea0421b8c74-kube-api-access-j9d4h\") pod \"collect-profiles-29330745-wbwkx\" (UID: \"d0a596b0-4162-4b10-8880-3ea0421b8c74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330745-wbwkx" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.899004 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e7167382-2267-4848-a423-e627253c38ed-trusted-ca-bundle\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.899071 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hqx5\" (UniqueName: \"kubernetes.io/projected/81d176a3-10ba-42f5-b716-d18a7dc1b5a8-kube-api-access-5hqx5\") pod \"openshift-apiserver-operator-796bbdcf4f-zg2sm\" (UID: \"81d176a3-10ba-42f5-b716-d18a7dc1b5a8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zg2sm" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.899117 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81d176a3-10ba-42f5-b716-d18a7dc1b5a8-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zg2sm\" (UID: \"81d176a3-10ba-42f5-b716-d18a7dc1b5a8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zg2sm" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.899144 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d6vt\" (UniqueName: \"kubernetes.io/projected/cb8de785-a298-414c-998c-7bd0585966e1-kube-api-access-6d6vt\") pod \"marketplace-operator-79b997595-xwlzm\" (UID: \"cb8de785-a298-414c-998c-7bd0585966e1\") " pod="openshift-marketplace/marketplace-operator-79b997595-xwlzm" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.899171 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e7167382-2267-4848-a423-e627253c38ed-encryption-config\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.899420 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/15866f74-4fd6-46df-b182-7cd58b5c1e0b-cert\") pod \"ingress-canary-7cvhs\" (UID: \"15866f74-4fd6-46df-b182-7cd58b5c1e0b\") " pod="openshift-ingress-canary/ingress-canary-7cvhs" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.899460 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/48528215-02d5-48aa-8289-9defecf47166-config-volume\") pod \"dns-default-rfcqs\" (UID: \"48528215-02d5-48aa-8289-9defecf47166\") " pod="openshift-dns/dns-default-rfcqs" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.899487 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54540da0-9e56-4ce2-bf77-1d055296e4b1-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rxt22\" (UID: \"54540da0-9e56-4ce2-bf77-1d055296e4b1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rxt22" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.899547 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/468622a9-56e9-47f6-aa19-b9df629881b7-signing-key\") pod \"service-ca-9c57cc56f-v5pq6\" (UID: \"468622a9-56e9-47f6-aa19-b9df629881b7\") " pod="openshift-service-ca/service-ca-9c57cc56f-v5pq6" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.899611 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pk9fk\" (UniqueName: \"kubernetes.io/projected/932c7fc9-8796-41bc-982f-74dfd0d887be-kube-api-access-pk9fk\") pod \"machine-config-controller-84d6567774-529ml\" (UID: \"932c7fc9-8796-41bc-982f-74dfd0d887be\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-529ml" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.899646 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/932c7fc9-8796-41bc-982f-74dfd0d887be-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-529ml\" (UID: \"932c7fc9-8796-41bc-982f-74dfd0d887be\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-529ml" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.899790 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/bd1bdacf-99ea-4afc-b804-2a87afb35cc7-profile-collector-cert\") pod \"olm-operator-6b444d44fb-ls9ng\" (UID: \"bd1bdacf-99ea-4afc-b804-2a87afb35cc7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ls9ng" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.899872 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2qlj\" (UniqueName: \"kubernetes.io/projected/54540da0-9e56-4ce2-bf77-1d055296e4b1-kube-api-access-v2qlj\") pod \"kube-storage-version-migrator-operator-b67b599dd-rxt22\" (UID: \"54540da0-9e56-4ce2-bf77-1d055296e4b1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rxt22" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.899896 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9kbr\" (UniqueName: \"kubernetes.io/projected/fd9598df-e972-4ee1-bfa4-e509f6bd04bc-kube-api-access-b9kbr\") pod \"machine-config-server-xn2j9\" (UID: \"fd9598df-e972-4ee1-bfa4-e509f6bd04bc\") " pod="openshift-machine-config-operator/machine-config-server-xn2j9" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.899915 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djztn\" (UniqueName: \"kubernetes.io/projected/bd1bdacf-99ea-4afc-b804-2a87afb35cc7-kube-api-access-djztn\") pod \"olm-operator-6b444d44fb-ls9ng\" (UID: \"bd1bdacf-99ea-4afc-b804-2a87afb35cc7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ls9ng" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.899949 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zt5s\" (UniqueName: \"kubernetes.io/projected/468622a9-56e9-47f6-aa19-b9df629881b7-kube-api-access-2zt5s\") pod \"service-ca-9c57cc56f-v5pq6\" (UID: \"468622a9-56e9-47f6-aa19-b9df629881b7\") " pod="openshift-service-ca/service-ca-9c57cc56f-v5pq6" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.899968 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0a596b0-4162-4b10-8880-3ea0421b8c74-config-volume\") pod \"collect-profiles-29330745-wbwkx\" (UID: \"d0a596b0-4162-4b10-8880-3ea0421b8c74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330745-wbwkx" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.899990 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxl5f\" (UniqueName: \"kubernetes.io/projected/8307e459-035b-4c33-bb8e-b5a698618036-kube-api-access-xxl5f\") pod \"downloads-7954f5f757-7bxtq\" (UID: \"8307e459-035b-4c33-bb8e-b5a698618036\") " pod="openshift-console/downloads-7954f5f757-7bxtq" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.900007 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e7167382-2267-4848-a423-e627253c38ed-audit\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.900024 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnhf6\" (UniqueName: \"kubernetes.io/projected/15866f74-4fd6-46df-b182-7cd58b5c1e0b-kube-api-access-dnhf6\") pod \"ingress-canary-7cvhs\" (UID: \"15866f74-4fd6-46df-b182-7cd58b5c1e0b\") " pod="openshift-ingress-canary/ingress-canary-7cvhs" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.900052 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbvl8\" (UniqueName: \"kubernetes.io/projected/07357227-cc26-43b8-89c7-4990980b2725-kube-api-access-hbvl8\") pod \"csi-hostpathplugin-fggsp\" (UID: \"07357227-cc26-43b8-89c7-4990980b2725\") " pod="hostpath-provisioner/csi-hostpathplugin-fggsp" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.900073 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e7167382-2267-4848-a423-e627253c38ed-etcd-client\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.900144 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3b0dd795-8b21-43a2-9e50-b0cb668ac22c-apiservice-cert\") pod \"packageserver-d55dfcdfc-5z54z\" (UID: \"3b0dd795-8b21-43a2-9e50-b0cb668ac22c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z54z" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.900166 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/cb8de785-a298-414c-998c-7bd0585966e1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xwlzm\" (UID: \"cb8de785-a298-414c-998c-7bd0585966e1\") " pod="openshift-marketplace/marketplace-operator-79b997595-xwlzm" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.900185 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c286e288-6157-4fff-bb4a-927e823d451a-trusted-ca\") pod \"ingress-operator-5b745b69d9-jj9rd\" (UID: \"c286e288-6157-4fff-bb4a-927e823d451a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jj9rd" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.900203 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/932c7fc9-8796-41bc-982f-74dfd0d887be-proxy-tls\") pod \"machine-config-controller-84d6567774-529ml\" (UID: \"932c7fc9-8796-41bc-982f-74dfd0d887be\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-529ml" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.900226 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdxq7\" (UniqueName: \"kubernetes.io/projected/c286e288-6157-4fff-bb4a-927e823d451a-kube-api-access-xdxq7\") pod \"ingress-operator-5b745b69d9-jj9rd\" (UID: \"c286e288-6157-4fff-bb4a-927e823d451a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jj9rd" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.900244 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/07357227-cc26-43b8-89c7-4990980b2725-socket-dir\") pod \"csi-hostpathplugin-fggsp\" (UID: \"07357227-cc26-43b8-89c7-4990980b2725\") " pod="hostpath-provisioner/csi-hostpathplugin-fggsp" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.900263 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5m9gl\" (UniqueName: \"kubernetes.io/projected/3b0dd795-8b21-43a2-9e50-b0cb668ac22c-kube-api-access-5m9gl\") pod \"packageserver-d55dfcdfc-5z54z\" (UID: \"3b0dd795-8b21-43a2-9e50-b0cb668ac22c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z54z" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.900285 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/bd1bdacf-99ea-4afc-b804-2a87afb35cc7-srv-cert\") pod \"olm-operator-6b444d44fb-ls9ng\" (UID: \"bd1bdacf-99ea-4afc-b804-2a87afb35cc7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ls9ng" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.900304 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/62481657-1b38-4d8d-81d5-665bd2574467-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-t8v4d\" (UID: \"62481657-1b38-4d8d-81d5-665bd2574467\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t8v4d" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.900321 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kljdw\" (UniqueName: \"kubernetes.io/projected/62481657-1b38-4d8d-81d5-665bd2574467-kube-api-access-kljdw\") pod \"cluster-image-registry-operator-dc59b4c8b-t8v4d\" (UID: \"62481657-1b38-4d8d-81d5-665bd2574467\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t8v4d" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.902064 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/fd9598df-e972-4ee1-bfa4-e509f6bd04bc-certs\") pod \"machine-config-server-xn2j9\" (UID: \"fd9598df-e972-4ee1-bfa4-e509f6bd04bc\") " pod="openshift-machine-config-operator/machine-config-server-xn2j9" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.902156 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/09af6fe0-f3ab-4223-9b45-05509e575ac4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-5qkt6\" (UID: \"09af6fe0-f3ab-4223-9b45-05509e575ac4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5qkt6" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.902245 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2843a931-4390-4dae-a736-911e2fc18bef-srv-cert\") pod \"catalog-operator-68c6474976-gfnhv\" (UID: \"2843a931-4390-4dae-a736-911e2fc18bef\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gfnhv" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.902845 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/fd9598df-e972-4ee1-bfa4-e509f6bd04bc-node-bootstrap-token\") pod \"machine-config-server-xn2j9\" (UID: \"fd9598df-e972-4ee1-bfa4-e509f6bd04bc\") " pod="openshift-machine-config-operator/machine-config-server-xn2j9" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.903572 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/932c7fc9-8796-41bc-982f-74dfd0d887be-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-529ml\" (UID: \"932c7fc9-8796-41bc-982f-74dfd0d887be\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-529ml" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.903712 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e7167382-2267-4848-a423-e627253c38ed-trusted-ca-bundle\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.904354 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81d176a3-10ba-42f5-b716-d18a7dc1b5a8-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zg2sm\" (UID: \"81d176a3-10ba-42f5-b716-d18a7dc1b5a8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zg2sm" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.904596 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e7167382-2267-4848-a423-e627253c38ed-image-import-ca\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.905657 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0a596b0-4162-4b10-8880-3ea0421b8c74-config-volume\") pod \"collect-profiles-29330745-wbwkx\" (UID: \"d0a596b0-4162-4b10-8880-3ea0421b8c74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330745-wbwkx" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.908014 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3218d84a-b760-4002-bf1f-79d02293ad93-metrics-tls\") pod \"dns-operator-744455d44c-cw2cl\" (UID: \"3218d84a-b760-4002-bf1f-79d02293ad93\") " pod="openshift-dns-operator/dns-operator-744455d44c-cw2cl" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.908569 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/62481657-1b38-4d8d-81d5-665bd2574467-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-t8v4d\" (UID: \"62481657-1b38-4d8d-81d5-665bd2574467\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t8v4d" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.910950 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54540da0-9e56-4ce2-bf77-1d055296e4b1-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rxt22\" (UID: \"54540da0-9e56-4ce2-bf77-1d055296e4b1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rxt22" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.911160 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e7167382-2267-4848-a423-e627253c38ed-audit\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.913664 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54540da0-9e56-4ce2-bf77-1d055296e4b1-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rxt22\" (UID: \"54540da0-9e56-4ce2-bf77-1d055296e4b1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rxt22" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.914642 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/62481657-1b38-4d8d-81d5-665bd2574467-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-t8v4d\" (UID: \"62481657-1b38-4d8d-81d5-665bd2574467\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t8v4d" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.917879 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/bd1bdacf-99ea-4afc-b804-2a87afb35cc7-srv-cert\") pod \"olm-operator-6b444d44fb-ls9ng\" (UID: \"bd1bdacf-99ea-4afc-b804-2a87afb35cc7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ls9ng" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.920559 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/cb8de785-a298-414c-998c-7bd0585966e1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xwlzm\" (UID: \"cb8de785-a298-414c-998c-7bd0585966e1\") " pod="openshift-marketplace/marketplace-operator-79b997595-xwlzm" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.920689 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q75c2\" (UniqueName: \"kubernetes.io/projected/f25b8da6-8185-4faa-b800-3deef3a9b06b-kube-api-access-q75c2\") pod \"openshift-config-operator-7777fb866f-xkp55\" (UID: \"f25b8da6-8185-4faa-b800-3deef3a9b06b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xkp55" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.920778 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c286e288-6157-4fff-bb4a-927e823d451a-trusted-ca\") pod \"ingress-operator-5b745b69d9-jj9rd\" (UID: \"c286e288-6157-4fff-bb4a-927e823d451a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jj9rd" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.921189 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3b0dd795-8b21-43a2-9e50-b0cb668ac22c-apiservice-cert\") pod \"packageserver-d55dfcdfc-5z54z\" (UID: \"3b0dd795-8b21-43a2-9e50-b0cb668ac22c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z54z" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.921954 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3b0dd795-8b21-43a2-9e50-b0cb668ac22c-webhook-cert\") pod \"packageserver-d55dfcdfc-5z54z\" (UID: \"3b0dd795-8b21-43a2-9e50-b0cb668ac22c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z54z" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.922443 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/468622a9-56e9-47f6-aa19-b9df629881b7-signing-key\") pod \"service-ca-9c57cc56f-v5pq6\" (UID: \"468622a9-56e9-47f6-aa19-b9df629881b7\") " pod="openshift-service-ca/service-ca-9c57cc56f-v5pq6" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.924723 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c286e288-6157-4fff-bb4a-927e823d451a-metrics-tls\") pod \"ingress-operator-5b745b69d9-jj9rd\" (UID: \"c286e288-6157-4fff-bb4a-927e823d451a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jj9rd" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.925278 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/15866f74-4fd6-46df-b182-7cd58b5c1e0b-cert\") pod \"ingress-canary-7cvhs\" (UID: \"15866f74-4fd6-46df-b182-7cd58b5c1e0b\") " pod="openshift-ingress-canary/ingress-canary-7cvhs" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.925690 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e7167382-2267-4848-a423-e627253c38ed-encryption-config\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.926185 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0a596b0-4162-4b10-8880-3ea0421b8c74-secret-volume\") pod \"collect-profiles-29330745-wbwkx\" (UID: \"d0a596b0-4162-4b10-8880-3ea0421b8c74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330745-wbwkx" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.927082 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/932c7fc9-8796-41bc-982f-74dfd0d887be-proxy-tls\") pod \"machine-config-controller-84d6567774-529ml\" (UID: \"932c7fc9-8796-41bc-982f-74dfd0d887be\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-529ml" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.945485 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e7167382-2267-4848-a423-e627253c38ed-etcd-client\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.947454 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/33deb07b-91b4-45f1-88d4-deed38a97f36-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-zwd57\" (UID: \"33deb07b-91b4-45f1-88d4-deed38a97f36\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zwd57" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.956459 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qr2pg"] Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.956607 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b748d317-5cb9-4fd3-a6bc-4a83d2f77d97-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lxbgb\" (UID: \"b748d317-5cb9-4fd3-a6bc-4a83d2f77d97\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lxbgb" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.964064 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v28wj\" (UniqueName: \"kubernetes.io/projected/b816c481-6891-479c-99db-95f6e4abbc05-kube-api-access-v28wj\") pod \"package-server-manager-789f6589d5-94zbw\" (UID: \"b816c481-6891-479c-99db-95f6e4abbc05\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-94zbw" Oct 07 13:47:39 crc kubenswrapper[4959]: I1007 13:47:39.992978 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmk25\" (UniqueName: \"kubernetes.io/projected/fe29debd-926a-4fd7-900c-b91a4d203436-kube-api-access-jmk25\") pod \"cluster-samples-operator-665b6dd947-2bh9c\" (UID: \"fe29debd-926a-4fd7-900c-b91a4d203436\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2bh9c" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.001940 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.002154 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/07357227-cc26-43b8-89c7-4990980b2725-mountpoint-dir\") pod \"csi-hostpathplugin-fggsp\" (UID: \"07357227-cc26-43b8-89c7-4990980b2725\") " pod="hostpath-provisioner/csi-hostpathplugin-fggsp" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.002201 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/48528215-02d5-48aa-8289-9defecf47166-metrics-tls\") pod \"dns-default-rfcqs\" (UID: \"48528215-02d5-48aa-8289-9defecf47166\") " pod="openshift-dns/dns-default-rfcqs" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.002250 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8qs8\" (UniqueName: \"kubernetes.io/projected/2843a931-4390-4dae-a736-911e2fc18bef-kube-api-access-k8qs8\") pod \"catalog-operator-68c6474976-gfnhv\" (UID: \"2843a931-4390-4dae-a736-911e2fc18bef\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gfnhv" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.002278 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/07357227-cc26-43b8-89c7-4990980b2725-csi-data-dir\") pod \"csi-hostpathplugin-fggsp\" (UID: \"07357227-cc26-43b8-89c7-4990980b2725\") " pod="hostpath-provisioner/csi-hostpathplugin-fggsp" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.002302 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7mcx\" (UniqueName: \"kubernetes.io/projected/48528215-02d5-48aa-8289-9defecf47166-kube-api-access-t7mcx\") pod \"dns-default-rfcqs\" (UID: \"48528215-02d5-48aa-8289-9defecf47166\") " pod="openshift-dns/dns-default-rfcqs" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.002406 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/48528215-02d5-48aa-8289-9defecf47166-config-volume\") pod \"dns-default-rfcqs\" (UID: \"48528215-02d5-48aa-8289-9defecf47166\") " pod="openshift-dns/dns-default-rfcqs" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.002626 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbvl8\" (UniqueName: \"kubernetes.io/projected/07357227-cc26-43b8-89c7-4990980b2725-kube-api-access-hbvl8\") pod \"csi-hostpathplugin-fggsp\" (UID: \"07357227-cc26-43b8-89c7-4990980b2725\") " pod="hostpath-provisioner/csi-hostpathplugin-fggsp" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.002658 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/07357227-cc26-43b8-89c7-4990980b2725-socket-dir\") pod \"csi-hostpathplugin-fggsp\" (UID: \"07357227-cc26-43b8-89c7-4990980b2725\") " pod="hostpath-provisioner/csi-hostpathplugin-fggsp" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.002707 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/07357227-cc26-43b8-89c7-4990980b2725-registration-dir\") pod \"csi-hostpathplugin-fggsp\" (UID: \"07357227-cc26-43b8-89c7-4990980b2725\") " pod="hostpath-provisioner/csi-hostpathplugin-fggsp" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.002744 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/07357227-cc26-43b8-89c7-4990980b2725-plugins-dir\") pod \"csi-hostpathplugin-fggsp\" (UID: \"07357227-cc26-43b8-89c7-4990980b2725\") " pod="hostpath-provisioner/csi-hostpathplugin-fggsp" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.002765 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2843a931-4390-4dae-a736-911e2fc18bef-profile-collector-cert\") pod \"catalog-operator-68c6474976-gfnhv\" (UID: \"2843a931-4390-4dae-a736-911e2fc18bef\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gfnhv" Oct 07 13:47:40 crc kubenswrapper[4959]: E1007 13:47:40.003458 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:40.50343862 +0000 UTC m=+142.586842945 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.003575 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/07357227-cc26-43b8-89c7-4990980b2725-mountpoint-dir\") pod \"csi-hostpathplugin-fggsp\" (UID: \"07357227-cc26-43b8-89c7-4990980b2725\") " pod="hostpath-provisioner/csi-hostpathplugin-fggsp" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.004745 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/07357227-cc26-43b8-89c7-4990980b2725-csi-data-dir\") pod \"csi-hostpathplugin-fggsp\" (UID: \"07357227-cc26-43b8-89c7-4990980b2725\") " pod="hostpath-provisioner/csi-hostpathplugin-fggsp" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.004867 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/07357227-cc26-43b8-89c7-4990980b2725-socket-dir\") pod \"csi-hostpathplugin-fggsp\" (UID: \"07357227-cc26-43b8-89c7-4990980b2725\") " pod="hostpath-provisioner/csi-hostpathplugin-fggsp" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.004875 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/07357227-cc26-43b8-89c7-4990980b2725-plugins-dir\") pod \"csi-hostpathplugin-fggsp\" (UID: \"07357227-cc26-43b8-89c7-4990980b2725\") " pod="hostpath-provisioner/csi-hostpathplugin-fggsp" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.005155 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/07357227-cc26-43b8-89c7-4990980b2725-registration-dir\") pod \"csi-hostpathplugin-fggsp\" (UID: \"07357227-cc26-43b8-89c7-4990980b2725\") " pod="hostpath-provisioner/csi-hostpathplugin-fggsp" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.005724 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrbcf\" (UniqueName: \"kubernetes.io/projected/dcc8e980-62b6-48fb-8339-fab83032bd0a-kube-api-access-qrbcf\") pod \"authentication-operator-69f744f599-82gdf\" (UID: \"dcc8e980-62b6-48fb-8339-fab83032bd0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-82gdf" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.005851 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-82gdf" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.005872 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2843a931-4390-4dae-a736-911e2fc18bef-profile-collector-cert\") pod \"catalog-operator-68c6474976-gfnhv\" (UID: \"2843a931-4390-4dae-a736-911e2fc18bef\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gfnhv" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.006702 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/48528215-02d5-48aa-8289-9defecf47166-metrics-tls\") pod \"dns-default-rfcqs\" (UID: \"48528215-02d5-48aa-8289-9defecf47166\") " pod="openshift-dns/dns-default-rfcqs" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.007011 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/48528215-02d5-48aa-8289-9defecf47166-config-volume\") pod \"dns-default-rfcqs\" (UID: \"48528215-02d5-48aa-8289-9defecf47166\") " pod="openshift-dns/dns-default-rfcqs" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.032074 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9dwg\" (UniqueName: \"kubernetes.io/projected/1d9c7700-8ba2-444e-a3f0-5c5c6dd05585-kube-api-access-d9dwg\") pod \"machine-api-operator-5694c8668f-dt2k4\" (UID: \"1d9c7700-8ba2-444e-a3f0-5c5c6dd05585\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dt2k4" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.043337 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mx6p7" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.045860 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8k48v"] Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.053851 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-phxkp"] Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.063770 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-dt2k4" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.075192 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/47a07011-2fae-40a2-9fb1-fec3461d2352-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-ddcbg\" (UID: \"47a07011-2fae-40a2-9fb1-fec3461d2352\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ddcbg" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.078355 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ddcbg" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.084572 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2bh9c" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.087594 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs2fh\" (UniqueName: \"kubernetes.io/projected/e7167382-2267-4848-a423-e627253c38ed-kube-api-access-rs2fh\") pod \"apiserver-76f77b778f-tv7sf\" (UID: \"e7167382-2267-4848-a423-e627253c38ed\") " pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.103174 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-pcvlp"] Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.104001 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:40 crc kubenswrapper[4959]: E1007 13:47:40.104534 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:40.604519242 +0000 UTC m=+142.687923567 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:40 crc kubenswrapper[4959]: W1007 13:47:40.107930 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e645a3d_0c4a_4e27_83ff_5bb01f1da515.slice/crio-e804467a1eacb77f33149e0f2d906300c0e2042884d6a87e336c82968e72b41c WatchSource:0}: Error finding container e804467a1eacb77f33149e0f2d906300c0e2042884d6a87e336c82968e72b41c: Status 404 returned error can't find the container with id e804467a1eacb77f33149e0f2d906300c0e2042884d6a87e336c82968e72b41c Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.109349 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7526c\" (UniqueName: \"kubernetes.io/projected/3218d84a-b760-4002-bf1f-79d02293ad93-kube-api-access-7526c\") pod \"dns-operator-744455d44c-cw2cl\" (UID: \"3218d84a-b760-4002-bf1f-79d02293ad93\") " pod="openshift-dns-operator/dns-operator-744455d44c-cw2cl" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.122079 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zwd57" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.128690 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/62481657-1b38-4d8d-81d5-665bd2574467-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-t8v4d\" (UID: \"62481657-1b38-4d8d-81d5-665bd2574467\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t8v4d" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.128817 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xkp55" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.135842 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lxbgb" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.154895 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kljdw\" (UniqueName: \"kubernetes.io/projected/62481657-1b38-4d8d-81d5-665bd2574467-kube-api-access-kljdw\") pod \"cluster-image-registry-operator-dc59b4c8b-t8v4d\" (UID: \"62481657-1b38-4d8d-81d5-665bd2574467\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t8v4d" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.157661 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-94zbw" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.171411 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-jdvrw"] Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.173662 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s95w5" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.190485 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2qlj\" (UniqueName: \"kubernetes.io/projected/54540da0-9e56-4ce2-bf77-1d055296e4b1-kube-api-access-v2qlj\") pod \"kube-storage-version-migrator-operator-b67b599dd-rxt22\" (UID: \"54540da0-9e56-4ce2-bf77-1d055296e4b1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rxt22" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.204718 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.204904 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t8v4d" Oct 07 13:47:40 crc kubenswrapper[4959]: E1007 13:47:40.205303 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:40.705279665 +0000 UTC m=+142.788683990 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.205689 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d6vt\" (UniqueName: \"kubernetes.io/projected/cb8de785-a298-414c-998c-7bd0585966e1-kube-api-access-6d6vt\") pod \"marketplace-operator-79b997595-xwlzm\" (UID: \"cb8de785-a298-414c-998c-7bd0585966e1\") " pod="openshift-marketplace/marketplace-operator-79b997595-xwlzm" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.212402 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6r8tg\" (UniqueName: \"kubernetes.io/projected/09af6fe0-f3ab-4223-9b45-05509e575ac4-kube-api-access-6r8tg\") pod \"multus-admission-controller-857f4d67dd-5qkt6\" (UID: \"09af6fe0-f3ab-4223-9b45-05509e575ac4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5qkt6" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.230025 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9d4h\" (UniqueName: \"kubernetes.io/projected/d0a596b0-4162-4b10-8880-3ea0421b8c74-kube-api-access-j9d4h\") pod \"collect-profiles-29330745-wbwkx\" (UID: \"d0a596b0-4162-4b10-8880-3ea0421b8c74\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330745-wbwkx" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.248248 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330745-wbwkx" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.248429 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-q9cfv"] Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.252754 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hqx5\" (UniqueName: \"kubernetes.io/projected/81d176a3-10ba-42f5-b716-d18a7dc1b5a8-kube-api-access-5hqx5\") pod \"openshift-apiserver-operator-796bbdcf4f-zg2sm\" (UID: \"81d176a3-10ba-42f5-b716-d18a7dc1b5a8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zg2sm" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.267726 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxl5f\" (UniqueName: \"kubernetes.io/projected/8307e459-035b-4c33-bb8e-b5a698618036-kube-api-access-xxl5f\") pod \"downloads-7954f5f757-7bxtq\" (UID: \"8307e459-035b-4c33-bb8e-b5a698618036\") " pod="openshift-console/downloads-7954f5f757-7bxtq" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.285804 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djztn\" (UniqueName: \"kubernetes.io/projected/bd1bdacf-99ea-4afc-b804-2a87afb35cc7-kube-api-access-djztn\") pod \"olm-operator-6b444d44fb-ls9ng\" (UID: \"bd1bdacf-99ea-4afc-b804-2a87afb35cc7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ls9ng" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.309885 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:40 crc kubenswrapper[4959]: E1007 13:47:40.310746 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:40.810731987 +0000 UTC m=+142.894136312 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.311471 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zt5s\" (UniqueName: \"kubernetes.io/projected/468622a9-56e9-47f6-aa19-b9df629881b7-kube-api-access-2zt5s\") pod \"service-ca-9c57cc56f-v5pq6\" (UID: \"468622a9-56e9-47f6-aa19-b9df629881b7\") " pod="openshift-service-ca/service-ca-9c57cc56f-v5pq6" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.312161 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zt5s\" (UniqueName: \"kubernetes.io/projected/468622a9-56e9-47f6-aa19-b9df629881b7-kube-api-access-2zt5s\") pod \"service-ca-9c57cc56f-v5pq6\" (UID: \"468622a9-56e9-47f6-aa19-b9df629881b7\") " pod="openshift-service-ca/service-ca-9c57cc56f-v5pq6" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.312630 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zt5s\" (UniqueName: \"kubernetes.io/projected/468622a9-56e9-47f6-aa19-b9df629881b7-kube-api-access-2zt5s\") pod \"service-ca-9c57cc56f-v5pq6\" (UID: \"468622a9-56e9-47f6-aa19-b9df629881b7\") " pod="openshift-service-ca/service-ca-9c57cc56f-v5pq6" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.324990 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-cw2cl" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.333359 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zg2sm" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.350915 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.350953 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdxq7\" (UniqueName: \"kubernetes.io/projected/c286e288-6157-4fff-bb4a-927e823d451a-kube-api-access-xdxq7\") pod \"ingress-operator-5b745b69d9-jj9rd\" (UID: \"c286e288-6157-4fff-bb4a-927e823d451a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jj9rd" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.357776 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnhf6\" (UniqueName: \"kubernetes.io/projected/15866f74-4fd6-46df-b182-7cd58b5c1e0b-kube-api-access-dnhf6\") pod \"ingress-canary-7cvhs\" (UID: \"15866f74-4fd6-46df-b182-7cd58b5c1e0b\") " pod="openshift-ingress-canary/ingress-canary-7cvhs" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.374006 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c286e288-6157-4fff-bb4a-927e823d451a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-jj9rd\" (UID: \"c286e288-6157-4fff-bb4a-927e823d451a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jj9rd" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.390745 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-7bxtq" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.404481 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pk9fk\" (UniqueName: \"kubernetes.io/projected/932c7fc9-8796-41bc-982f-74dfd0d887be-kube-api-access-pk9fk\") pod \"machine-config-controller-84d6567774-529ml\" (UID: \"932c7fc9-8796-41bc-982f-74dfd0d887be\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-529ml" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.406262 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5m9gl\" (UniqueName: \"kubernetes.io/projected/3b0dd795-8b21-43a2-9e50-b0cb668ac22c-kube-api-access-5m9gl\") pod \"packageserver-d55dfcdfc-5z54z\" (UID: \"3b0dd795-8b21-43a2-9e50-b0cb668ac22c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z54z" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.415370 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:40 crc kubenswrapper[4959]: E1007 13:47:40.415539 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:40.915513511 +0000 UTC m=+142.998917836 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.417088 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:40 crc kubenswrapper[4959]: E1007 13:47:40.417464 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:40.917454525 +0000 UTC m=+143.000858850 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.426417 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8k48v" event={"ID":"5d295e40-2e2f-419e-9db4-cee7392d913e","Type":"ContainerStarted","Data":"eb1a2b6b840454f1a0e52f9ac77a75728ac160a9140d296ccb4036eb528a4932"} Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.430588 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-q9cfv" event={"ID":"9872926d-9b2f-489a-804d-749ec1e2c136","Type":"ContainerStarted","Data":"d3f06be5f549da8ff9d9b7a1d473e6e17111ed78d1fe800a569dfa6e53f32eba"} Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.434956 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-qr2pg" event={"ID":"dd2f3a74-d8f7-4970-b748-39036ccd4aab","Type":"ContainerStarted","Data":"5f47ef5d89314e017bb25d500f9f6bb79a2af3cce59587e13cfef5a427f70670"} Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.448616 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87" event={"ID":"0ff428ab-6c11-4571-9717-9ac0bfe462d4","Type":"ContainerStarted","Data":"91dc56d48a2de7414ced1f338d667518a1ffdd206b127332624a0f63607898dc"} Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.449126 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.449397 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rxt22" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.454434 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9kbr\" (UniqueName: \"kubernetes.io/projected/fd9598df-e972-4ee1-bfa4-e509f6bd04bc-kube-api-access-b9kbr\") pod \"machine-config-server-xn2j9\" (UID: \"fd9598df-e972-4ee1-bfa4-e509f6bd04bc\") " pod="openshift-machine-config-operator/machine-config-server-xn2j9" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.457664 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7mcx\" (UniqueName: \"kubernetes.io/projected/48528215-02d5-48aa-8289-9defecf47166-kube-api-access-t7mcx\") pod \"dns-default-rfcqs\" (UID: \"48528215-02d5-48aa-8289-9defecf47166\") " pod="openshift-dns/dns-default-rfcqs" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.479404 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-phxkp" event={"ID":"2e645a3d-0c4a-4e27-83ff-5bb01f1da515","Type":"ContainerStarted","Data":"e804467a1eacb77f33149e0f2d906300c0e2042884d6a87e336c82968e72b41c"} Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.481614 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8qs8\" (UniqueName: \"kubernetes.io/projected/2843a931-4390-4dae-a736-911e2fc18bef-kube-api-access-k8qs8\") pod \"catalog-operator-68c6474976-gfnhv\" (UID: \"2843a931-4390-4dae-a736-911e2fc18bef\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gfnhv" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.487445 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xwlzm" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.495709 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jj9rd" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.496733 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbvl8\" (UniqueName: \"kubernetes.io/projected/07357227-cc26-43b8-89c7-4990980b2725-kube-api-access-hbvl8\") pod \"csi-hostpathplugin-fggsp\" (UID: \"07357227-cc26-43b8-89c7-4990980b2725\") " pod="hostpath-provisioner/csi-hostpathplugin-fggsp" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.498307 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-5qkt6" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.509065 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-pcvlp" event={"ID":"4fca2a0e-545f-43b0-b4ec-e5f15babcd71","Type":"ContainerStarted","Data":"56a8e024537dcef4e292424b730f2cf4efd6fb405e8b6b483bf8262bf5e4f1d4"} Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.522621 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l78hd" event={"ID":"bb4ba104-3024-4e76-a15c-77decd3a67c6","Type":"ContainerStarted","Data":"aac080498a91d1be838cbf67af3b25b6eccfbf175817941bba854dad749c9413"} Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.528397 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-529ml" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.529071 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ls9ng" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.529334 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gfnhv" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.529888 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:40 crc kubenswrapper[4959]: E1007 13:47:40.532006 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:41.031988788 +0000 UTC m=+143.115393113 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.540272 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jdvrw" event={"ID":"6158d485-779d-4fe3-8772-32b33a67b1f0","Type":"ContainerStarted","Data":"1219c1310c9d8dddddc16d97719f33d3833253f254588e624dd4cb8186a17eb0"} Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.540547 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z54z" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.557893 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-v5pq6" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.569633 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-rfcqs" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.572801 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-7cvhs" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.600435 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-fggsp" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.602275 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-82gdf"] Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.608556 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-xn2j9" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.617456 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-8nt7g" event={"ID":"67543fb5-5742-4067-a953-7d94ac45e415","Type":"ContainerStarted","Data":"7b3b2cdf16db99bd97d72627b0d5f0faf1233f58a979e3239d39633a06b811e9"} Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.617492 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-8nt7g" event={"ID":"67543fb5-5742-4067-a953-7d94ac45e415","Type":"ContainerStarted","Data":"7d6acde377fd5cefe74434e0c5ff5f57d3df52b0a6522fe95bc44649a2c6d99f"} Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.653113 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" event={"ID":"cc0f9f33-98c7-433b-a2e8-2a1dd0066c97","Type":"ContainerStarted","Data":"b6618bf546e72f3c60743e3e920f75cb638dea00f14a65519d0430db38169d74"} Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.694191 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:40 crc kubenswrapper[4959]: E1007 13:47:40.695811 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:41.195798142 +0000 UTC m=+143.279202467 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.744148 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-9drmk" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.744194 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.744208 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mx6p7"] Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.744225 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2bh9c"] Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.755926 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-dt2k4"] Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.760943 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ddcbg"] Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.808136 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:40 crc kubenswrapper[4959]: E1007 13:47:40.810422 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:41.310399867 +0000 UTC m=+143.393804192 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.812805 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-8nt7g" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.875470 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.879843 4959 patch_prober.go:28] interesting pod/router-default-5444994796-8nt7g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 13:47:40 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 07 13:47:40 crc kubenswrapper[4959]: [+]process-running ok Oct 07 13:47:40 crc kubenswrapper[4959]: healthz check failed Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.880141 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nt7g" podUID="67543fb5-5742-4067-a953-7d94ac45e415" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.911947 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:40 crc kubenswrapper[4959]: E1007 13:47:40.912897 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:41.412880347 +0000 UTC m=+143.496284672 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:40 crc kubenswrapper[4959]: I1007 13:47:40.982268 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-s95w5"] Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.013691 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:41 crc kubenswrapper[4959]: E1007 13:47:41.014254 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:41.514236956 +0000 UTC m=+143.597641281 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.041193 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-xkp55"] Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.125787 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:41 crc kubenswrapper[4959]: E1007 13:47:41.126083 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:41.626068905 +0000 UTC m=+143.709473230 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.244926 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:41 crc kubenswrapper[4959]: E1007 13:47:41.245812 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:41.745795442 +0000 UTC m=+143.829199767 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.347485 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:41 crc kubenswrapper[4959]: E1007 13:47:41.348050 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:41.848032465 +0000 UTC m=+143.931436790 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.423504 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zwd57"] Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.448817 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:41 crc kubenswrapper[4959]: E1007 13:47:41.449033 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:41.949014314 +0000 UTC m=+144.032418639 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.449555 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:41 crc kubenswrapper[4959]: E1007 13:47:41.450013 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:41.950003622 +0000 UTC m=+144.033407947 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.550557 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:41 crc kubenswrapper[4959]: E1007 13:47:41.550947 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:42.050921739 +0000 UTC m=+144.134326064 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.551263 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:41 crc kubenswrapper[4959]: E1007 13:47:41.551644 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:42.051632788 +0000 UTC m=+144.135037113 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.648901 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" podStartSLOduration=118.648884594 podStartE2EDuration="1m58.648884594s" podCreationTimestamp="2025-10-07 13:45:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:41.643525516 +0000 UTC m=+143.726929831" watchObservedRunningTime="2025-10-07 13:47:41.648884594 +0000 UTC m=+143.732288919" Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.654738 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:41 crc kubenswrapper[4959]: E1007 13:47:41.655908 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:42.155888028 +0000 UTC m=+144.239292353 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.669729 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-dt2k4" event={"ID":"1d9c7700-8ba2-444e-a3f0-5c5c6dd05585","Type":"ContainerStarted","Data":"75824a5437c535d1e5c6d0232feab2330d9ff940ab190e53da05289030269afa"} Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.683732 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8k48v" event={"ID":"5d295e40-2e2f-419e-9db4-cee7392d913e","Type":"ContainerStarted","Data":"0c5ba3acceb28dd4e9f7efe1028f06f9cb6cc774cc8c06e684dbaff0f8da33ad"} Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.688025 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mx6p7" event={"ID":"4a2a7f86-33ef-42d0-93b2-765b74d5c629","Type":"ContainerStarted","Data":"251de6db04001d9807089d01b2bebf291fb8730caef2abb1474e5eb226e9953a"} Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.688060 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mx6p7" event={"ID":"4a2a7f86-33ef-42d0-93b2-765b74d5c629","Type":"ContainerStarted","Data":"f1fa06713dc30e1b47420035189dbf0a30d39fd52e06857e5d83cc47b55d9997"} Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.691150 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2bh9c" event={"ID":"fe29debd-926a-4fd7-900c-b91a4d203436","Type":"ContainerStarted","Data":"f04a2e8d28ae80286a37c3f82720e1055ee28114fb325372881761ee339298cb"} Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.705191 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-qr2pg" event={"ID":"dd2f3a74-d8f7-4970-b748-39036ccd4aab","Type":"ContainerStarted","Data":"916d12bfbdd20d3482928c53413214e783530c16e7ac08831b946df3f52d2e86"} Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.711854 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-phxkp" event={"ID":"2e645a3d-0c4a-4e27-83ff-5bb01f1da515","Type":"ContainerStarted","Data":"70b561d1f96729bdc101213af122d5d6e797ac2f102c7c5c5885fb5cc03853d8"} Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.714292 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-phxkp" Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.715647 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" podStartSLOduration=117.715625018 podStartE2EDuration="1m57.715625018s" podCreationTimestamp="2025-10-07 13:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:41.714693162 +0000 UTC m=+143.798097487" watchObservedRunningTime="2025-10-07 13:47:41.715625018 +0000 UTC m=+143.799029343" Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.720248 4959 patch_prober.go:28] interesting pod/console-operator-58897d9998-phxkp container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.720318 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-phxkp" podUID="2e645a3d-0c4a-4e27-83ff-5bb01f1da515" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.726050 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-xn2j9" event={"ID":"fd9598df-e972-4ee1-bfa4-e509f6bd04bc","Type":"ContainerStarted","Data":"85a8a0389b4470fd7ad64a77f88f3290abe99184222d3662ca8857fd90ad7b69"} Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.726111 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-xn2j9" event={"ID":"fd9598df-e972-4ee1-bfa4-e509f6bd04bc","Type":"ContainerStarted","Data":"ad5c6f06aeee32e88f799f00d3eb08b54d30ae86ff51910c8de8e15c8b4f981c"} Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.733684 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-pcvlp" event={"ID":"4fca2a0e-545f-43b0-b4ec-e5f15babcd71","Type":"ContainerStarted","Data":"d15871c5e854d27c40c25e1eca507e49d5a98e2edd4342505bbcf6c1517412fd"} Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.734576 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xkp55" event={"ID":"f25b8da6-8185-4faa-b800-3deef3a9b06b","Type":"ContainerStarted","Data":"108e5567c624ecdcc043611c460c59b113ba0af00aa8229b291de201d9affe9a"} Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.746573 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-9drmk" podStartSLOduration=118.746554742 podStartE2EDuration="1m58.746554742s" podCreationTimestamp="2025-10-07 13:45:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:41.742897741 +0000 UTC m=+143.826302066" watchObservedRunningTime="2025-10-07 13:47:41.746554742 +0000 UTC m=+143.829959067" Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.754572 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-82gdf" event={"ID":"dcc8e980-62b6-48fb-8339-fab83032bd0a","Type":"ContainerStarted","Data":"c9e1364d200c9393ab81f7d7ed53c34f5b4efd188301e8335dc2ef7bfcfe5c3a"} Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.754637 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-82gdf" event={"ID":"dcc8e980-62b6-48fb-8339-fab83032bd0a","Type":"ContainerStarted","Data":"eb7af48a3971023458c2b814b833ad9f2ef2194dc43a8406a9555a7da2caea92"} Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.757203 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:41 crc kubenswrapper[4959]: E1007 13:47:41.764692 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:42.264673722 +0000 UTC m=+144.348078267 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.794113 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jdvrw" event={"ID":"6158d485-779d-4fe3-8772-32b33a67b1f0","Type":"ContainerStarted","Data":"219e00d4cbfb181d093311f235584bcdfab78f80518e41b5d04f84aa81ea7df6"} Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.817378 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l78hd" podStartSLOduration=118.817358247 podStartE2EDuration="1m58.817358247s" podCreationTimestamp="2025-10-07 13:45:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:41.816572326 +0000 UTC m=+143.899976651" watchObservedRunningTime="2025-10-07 13:47:41.817358247 +0000 UTC m=+143.900762572" Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.829597 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-q9cfv" event={"ID":"9872926d-9b2f-489a-804d-749ec1e2c136","Type":"ContainerStarted","Data":"79e92b73d680ec7dc90734a553058a431ae660efb8db60a23145c1ecaa55fcbb"} Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.836140 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s95w5" event={"ID":"de5396bf-4d94-46be-87e1-7b587147a6a4","Type":"ContainerStarted","Data":"6fa78ad51231a87082c634a11bf97c9efe1d0b77442481e3d2abb11773a69ae0"} Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.842544 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zwd57" event={"ID":"33deb07b-91b4-45f1-88d4-deed38a97f36","Type":"ContainerStarted","Data":"5c55f10adee64822fc9f80c395b46783b05fb0ec1f51f003c48bc5ce066179fa"} Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.845140 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ddcbg" event={"ID":"47a07011-2fae-40a2-9fb1-fec3461d2352","Type":"ContainerStarted","Data":"3f342093ebe40f536a404aa81bcb642c7a17c458bf05c757640fc8dc190aed3e"} Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.845277 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-8nt7g" podStartSLOduration=117.845259878 podStartE2EDuration="1m57.845259878s" podCreationTimestamp="2025-10-07 13:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:41.841212546 +0000 UTC m=+143.924616871" watchObservedRunningTime="2025-10-07 13:47:41.845259878 +0000 UTC m=+143.928664203" Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.858460 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:41 crc kubenswrapper[4959]: E1007 13:47:41.858647 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:42.358619777 +0000 UTC m=+144.442024102 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:41 crc kubenswrapper[4959]: I1007 13:47:41.961116 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:41 crc kubenswrapper[4959]: E1007 13:47:41.965069 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:42.465047676 +0000 UTC m=+144.548452001 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.013324 4959 patch_prober.go:28] interesting pod/router-default-5444994796-8nt7g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 13:47:42 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 07 13:47:42 crc kubenswrapper[4959]: [+]process-running ok Oct 07 13:47:42 crc kubenswrapper[4959]: healthz check failed Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.013388 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nt7g" podUID="67543fb5-5742-4067-a953-7d94ac45e415" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.063076 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:42 crc kubenswrapper[4959]: E1007 13:47:42.063393 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:42.563363692 +0000 UTC m=+144.646768067 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.063529 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:42 crc kubenswrapper[4959]: E1007 13:47:42.063931 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:42.563920667 +0000 UTC m=+144.647324992 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.078373 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87" podStartSLOduration=118.078356456 podStartE2EDuration="1m58.078356456s" podCreationTimestamp="2025-10-07 13:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:42.077524643 +0000 UTC m=+144.160928968" watchObservedRunningTime="2025-10-07 13:47:42.078356456 +0000 UTC m=+144.161760781" Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.164800 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:42 crc kubenswrapper[4959]: E1007 13:47:42.165186 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:42.665166563 +0000 UTC m=+144.748570888 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.264663 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-q9cfv" podStartSLOduration=118.264643631 podStartE2EDuration="1m58.264643631s" podCreationTimestamp="2025-10-07 13:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:42.22839652 +0000 UTC m=+144.311800845" watchObservedRunningTime="2025-10-07 13:47:42.264643631 +0000 UTC m=+144.348047956" Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.265801 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mx6p7" podStartSLOduration=119.265796993 podStartE2EDuration="1m59.265796993s" podCreationTimestamp="2025-10-07 13:45:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:42.264493767 +0000 UTC m=+144.347898092" watchObservedRunningTime="2025-10-07 13:47:42.265796993 +0000 UTC m=+144.349201318" Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.266649 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:42 crc kubenswrapper[4959]: E1007 13:47:42.266900 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:42.766890433 +0000 UTC m=+144.850294758 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.298402 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-94zbw"] Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.324617 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330745-wbwkx"] Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.351291 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-xn2j9" podStartSLOduration=5.351272393 podStartE2EDuration="5.351272393s" podCreationTimestamp="2025-10-07 13:47:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:42.346331807 +0000 UTC m=+144.429736132" watchObservedRunningTime="2025-10-07 13:47:42.351272393 +0000 UTC m=+144.434676718" Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.355330 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lxbgb"] Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.367408 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.367815 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t8v4d"] Oct 07 13:47:42 crc kubenswrapper[4959]: E1007 13:47:42.367931 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:42.867916143 +0000 UTC m=+144.951320458 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.398988 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-qr2pg" podStartSLOduration=118.398969301 podStartE2EDuration="1m58.398969301s" podCreationTimestamp="2025-10-07 13:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:42.388520142 +0000 UTC m=+144.471924467" watchObservedRunningTime="2025-10-07 13:47:42.398969301 +0000 UTC m=+144.482373626" Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.412435 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zg2sm"] Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.469284 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-phxkp" podStartSLOduration=119.469266872 podStartE2EDuration="1m59.469266872s" podCreationTimestamp="2025-10-07 13:45:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:42.423775326 +0000 UTC m=+144.507179651" watchObservedRunningTime="2025-10-07 13:47:42.469266872 +0000 UTC m=+144.552671197" Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.473938 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:42 crc kubenswrapper[4959]: E1007 13:47:42.474366 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:42.974351513 +0000 UTC m=+145.057755838 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.491208 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rxt22"] Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.523943 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-jj9rd"] Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.525908 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gfnhv"] Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.531116 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-82gdf" podStartSLOduration=119.531075409 podStartE2EDuration="1m59.531075409s" podCreationTimestamp="2025-10-07 13:45:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:42.514714737 +0000 UTC m=+144.598119072" watchObservedRunningTime="2025-10-07 13:47:42.531075409 +0000 UTC m=+144.614479734" Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.563778 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-7bxtq"] Oct 07 13:47:42 crc kubenswrapper[4959]: W1007 13:47:42.569524 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54540da0_9e56_4ce2_bf77_1d055296e4b1.slice/crio-d9741983354eefe34ce5fe1f8622ba84ce6bacb9aaa31be157f1e919484a15f6 WatchSource:0}: Error finding container d9741983354eefe34ce5fe1f8622ba84ce6bacb9aaa31be157f1e919484a15f6: Status 404 returned error can't find the container with id d9741983354eefe34ce5fe1f8622ba84ce6bacb9aaa31be157f1e919484a15f6 Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.590968 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:42 crc kubenswrapper[4959]: E1007 13:47:42.591568 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:43.091553429 +0000 UTC m=+145.174957744 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.607722 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8k48v" podStartSLOduration=118.607695085 podStartE2EDuration="1m58.607695085s" podCreationTimestamp="2025-10-07 13:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:42.590562512 +0000 UTC m=+144.673966847" watchObservedRunningTime="2025-10-07 13:47:42.607695085 +0000 UTC m=+144.691099410" Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.613830 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-tv7sf"] Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.637809 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z54z"] Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.691888 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:42 crc kubenswrapper[4959]: E1007 13:47:42.692283 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:43.192272961 +0000 UTC m=+145.275677286 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.702975 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-pcvlp" podStartSLOduration=119.702938856 podStartE2EDuration="1m59.702938856s" podCreationTimestamp="2025-10-07 13:45:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:42.689241267 +0000 UTC m=+144.772645592" watchObservedRunningTime="2025-10-07 13:47:42.702938856 +0000 UTC m=+144.786343181" Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.703872 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-7cvhs"] Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.713340 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-rfcqs"] Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.714580 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-fggsp"] Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.714596 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-5qkt6"] Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.787631 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-cw2cl"] Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.805055 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-529ml"] Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.806865 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-v5pq6"] Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.807093 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:42 crc kubenswrapper[4959]: E1007 13:47:42.807228 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:43.307201545 +0000 UTC m=+145.390605870 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.808792 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:42 crc kubenswrapper[4959]: E1007 13:47:42.810082 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:43.310062214 +0000 UTC m=+145.393466539 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.815364 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ls9ng"] Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.824164 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xwlzm"] Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.815646 4959 patch_prober.go:28] interesting pod/router-default-5444994796-8nt7g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 13:47:42 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 07 13:47:42 crc kubenswrapper[4959]: [+]process-running ok Oct 07 13:47:42 crc kubenswrapper[4959]: healthz check failed Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.824239 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nt7g" podUID="67543fb5-5742-4067-a953-7d94ac45e415" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.853425 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rfcqs" event={"ID":"48528215-02d5-48aa-8289-9defecf47166","Type":"ContainerStarted","Data":"a54a9f2a7fe99a86b3d1e9487b5bab2034bc9ca697cd9cb4183e4341afa9fcee"} Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.857591 4959 generic.go:334] "Generic (PLEG): container finished" podID="f25b8da6-8185-4faa-b800-3deef3a9b06b" containerID="a2c589ffc575664a5dd1198201ce574f9b322dcf9f2bd0e5ddf66fac672bd2d3" exitCode=0 Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.857653 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xkp55" event={"ID":"f25b8da6-8185-4faa-b800-3deef3a9b06b","Type":"ContainerDied","Data":"a2c589ffc575664a5dd1198201ce574f9b322dcf9f2bd0e5ddf66fac672bd2d3"} Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.868636 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z54z" event={"ID":"3b0dd795-8b21-43a2-9e50-b0cb668ac22c","Type":"ContainerStarted","Data":"062403a6e642a2616de159d95bebadf9bfba25c9a7a76c1ecc58844dfd4ac73c"} Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.874787 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zwd57" event={"ID":"33deb07b-91b4-45f1-88d4-deed38a97f36","Type":"ContainerStarted","Data":"7da8dfad44c3f818429f8e4aafdc2955a82fcebb4ca87fda8fca38e8004bb792"} Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.909431 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:42 crc kubenswrapper[4959]: E1007 13:47:42.911527 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:43.411493946 +0000 UTC m=+145.494898481 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.912256 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jdvrw" event={"ID":"6158d485-779d-4fe3-8772-32b33a67b1f0","Type":"ContainerStarted","Data":"a68578316a54336f3fc1a3b941881bf7b83a41b675847027a5cab36c7bb950d8"} Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.922155 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ddcbg" event={"ID":"47a07011-2fae-40a2-9fb1-fec3461d2352","Type":"ContainerStarted","Data":"c38d93e417d98e049cdf95f683fddb58d66a7153b352a00fac26bc0f33fd8599"} Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.957383 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jj9rd" event={"ID":"c286e288-6157-4fff-bb4a-927e823d451a","Type":"ContainerStarted","Data":"602c24cec6db49d43f8574ff9f4c930ef346da9520bfbb4cc8c6bde4ef17f55e"} Oct 07 13:47:42 crc kubenswrapper[4959]: I1007 13:47:42.959248 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jdvrw" podStartSLOduration=118.959219104 podStartE2EDuration="1m58.959219104s" podCreationTimestamp="2025-10-07 13:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:42.956951111 +0000 UTC m=+145.040355436" watchObservedRunningTime="2025-10-07 13:47:42.959219104 +0000 UTC m=+145.042623429" Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.008343 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gfnhv" event={"ID":"2843a931-4390-4dae-a736-911e2fc18bef","Type":"ContainerStarted","Data":"00703286585b096aad5c14d785dd0df63904618b097318db0e220f12edfe0ee5"} Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.011267 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.012576 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zwd57" podStartSLOduration=119.012557467 podStartE2EDuration="1m59.012557467s" podCreationTimestamp="2025-10-07 13:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:43.010422718 +0000 UTC m=+145.093827043" watchObservedRunningTime="2025-10-07 13:47:43.012557467 +0000 UTC m=+145.095961792" Oct 07 13:47:43 crc kubenswrapper[4959]: E1007 13:47:43.012973 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:43.512959818 +0000 UTC m=+145.596364143 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.034981 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-dt2k4" event={"ID":"1d9c7700-8ba2-444e-a3f0-5c5c6dd05585","Type":"ContainerStarted","Data":"fe52eece882000b27d220dba2ca8bb66e9c1274fea0e27e7870b06001cce099f"} Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.035159 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-dt2k4" event={"ID":"1d9c7700-8ba2-444e-a3f0-5c5c6dd05585","Type":"ContainerStarted","Data":"947ad028e098c0e72e8c818d6675e3508961113299b1d62dd312000441e645bf"} Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.051192 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ddcbg" podStartSLOduration=119.051091081 podStartE2EDuration="1m59.051091081s" podCreationTimestamp="2025-10-07 13:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:43.05068772 +0000 UTC m=+145.134092045" watchObservedRunningTime="2025-10-07 13:47:43.051091081 +0000 UTC m=+145.134495406" Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.052583 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-7cvhs" event={"ID":"15866f74-4fd6-46df-b182-7cd58b5c1e0b","Type":"ContainerStarted","Data":"a49932a9f880bba7e61fdefddb64a25fbf7d555d0aa9449962d455da0f21d115"} Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.074855 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2bh9c" event={"ID":"fe29debd-926a-4fd7-900c-b91a4d203436","Type":"ContainerStarted","Data":"0626c8dadb03829eb3d5eee7ddd4e42e740bf588c2210fbdfcbf4e83f1fab9ea"} Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.076363 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lxbgb" event={"ID":"b748d317-5cb9-4fd3-a6bc-4a83d2f77d97","Type":"ContainerStarted","Data":"fda4c711a725333e8fc754456f8fc098ea68debad65232094bebe202e7510999"} Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.077156 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rxt22" event={"ID":"54540da0-9e56-4ce2-bf77-1d055296e4b1","Type":"ContainerStarted","Data":"d9741983354eefe34ce5fe1f8622ba84ce6bacb9aaa31be157f1e919484a15f6"} Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.101739 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-94zbw" event={"ID":"b816c481-6891-479c-99db-95f6e4abbc05","Type":"ContainerStarted","Data":"41fe53556604118ca2e7cd02a13ce7aeb10b52990d5f66466a85cce6a2a50bbe"} Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.101780 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-94zbw" event={"ID":"b816c481-6891-479c-99db-95f6e4abbc05","Type":"ContainerStarted","Data":"340a8aacc0710c6fb4ce7788aa485d44481e4e48ffb02dde009e35d42878f682"} Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.108928 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330745-wbwkx" event={"ID":"d0a596b0-4162-4b10-8880-3ea0421b8c74","Type":"ContainerStarted","Data":"a7017d01f13370140f49dfc5cb484478c43956cb32641304263fe6937e7d534b"} Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.108978 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330745-wbwkx" event={"ID":"d0a596b0-4162-4b10-8880-3ea0421b8c74","Type":"ContainerStarted","Data":"1e512e40764b77383cb5bbfdec10776f2a2dca9062b7365ad69d76968d5aa5d3"} Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.115266 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:43 crc kubenswrapper[4959]: E1007 13:47:43.117686 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:43.61766845 +0000 UTC m=+145.701072765 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.119260 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-dt2k4" podStartSLOduration=119.119245974 podStartE2EDuration="1m59.119245974s" podCreationTimestamp="2025-10-07 13:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:43.11693273 +0000 UTC m=+145.200337055" watchObservedRunningTime="2025-10-07 13:47:43.119245974 +0000 UTC m=+145.202650299" Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.139241 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" event={"ID":"e7167382-2267-4848-a423-e627253c38ed","Type":"ContainerStarted","Data":"e1bf870e4dc8be4387abf3c0fe67c7bb5d537dff3ab5c48dc32f7fd3ea9f5752"} Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.143274 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29330745-wbwkx" podStartSLOduration=120.143257687 podStartE2EDuration="2m0.143257687s" podCreationTimestamp="2025-10-07 13:45:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:43.141587331 +0000 UTC m=+145.224991656" watchObservedRunningTime="2025-10-07 13:47:43.143257687 +0000 UTC m=+145.226662012" Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.200059 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s95w5" event={"ID":"de5396bf-4d94-46be-87e1-7b587147a6a4","Type":"ContainerStarted","Data":"92ff10cfd9a21b949575315186a95257f2930e3aa27be4b86663a4cd3a2bca0c"} Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.211217 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-7bxtq" event={"ID":"8307e459-035b-4c33-bb8e-b5a698618036","Type":"ContainerStarted","Data":"e8d627190597ecdc999d8c6ebcf2913dd4b7c5a0eff48bcf43b2fdd7a7a07ba3"} Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.226649 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t8v4d" event={"ID":"62481657-1b38-4d8d-81d5-665bd2574467","Type":"ContainerStarted","Data":"af0f1e1d80756504f83996623913b7089423a8b193df96f3226ae877bca18d41"} Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.228461 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:43 crc kubenswrapper[4959]: E1007 13:47:43.228861 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:43.728847081 +0000 UTC m=+145.812251406 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.233953 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zg2sm" event={"ID":"81d176a3-10ba-42f5-b716-d18a7dc1b5a8","Type":"ContainerStarted","Data":"9189a77eb04c816361f0ee5085502626f492a70691f9c56db5d3dfe9308ff4c7"} Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.263784 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s95w5" podStartSLOduration=119.263763585 podStartE2EDuration="1m59.263763585s" podCreationTimestamp="2025-10-07 13:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:43.227778821 +0000 UTC m=+145.311183146" watchObservedRunningTime="2025-10-07 13:47:43.263763585 +0000 UTC m=+145.347167920" Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.264742 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t8v4d" podStartSLOduration=119.264735462 podStartE2EDuration="1m59.264735462s" podCreationTimestamp="2025-10-07 13:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:43.262978493 +0000 UTC m=+145.346382818" watchObservedRunningTime="2025-10-07 13:47:43.264735462 +0000 UTC m=+145.348139787" Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.336040 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:43 crc kubenswrapper[4959]: E1007 13:47:43.336839 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:43.836792262 +0000 UTC m=+145.920196577 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.337856 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:43 crc kubenswrapper[4959]: E1007 13:47:43.363386 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:43.863336305 +0000 UTC m=+145.946740630 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.439980 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:43 crc kubenswrapper[4959]: E1007 13:47:43.440518 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:43.940493456 +0000 UTC m=+146.023897781 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.549929 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:43 crc kubenswrapper[4959]: E1007 13:47:43.550881 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:44.050863834 +0000 UTC m=+146.134268169 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.652047 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:43 crc kubenswrapper[4959]: E1007 13:47:43.652243 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:44.152207653 +0000 UTC m=+146.235611978 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.652527 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:43 crc kubenswrapper[4959]: E1007 13:47:43.652936 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:44.152920003 +0000 UTC m=+146.236324318 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.776422 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:43 crc kubenswrapper[4959]: E1007 13:47:43.777268 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:44.277250567 +0000 UTC m=+146.360654902 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.784981 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-phxkp" Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.803886 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zg2sm" podStartSLOduration=120.803863982 podStartE2EDuration="2m0.803863982s" podCreationTimestamp="2025-10-07 13:45:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:43.302239638 +0000 UTC m=+145.385643963" watchObservedRunningTime="2025-10-07 13:47:43.803863982 +0000 UTC m=+145.887268307" Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.826409 4959 patch_prober.go:28] interesting pod/router-default-5444994796-8nt7g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 13:47:43 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 07 13:47:43 crc kubenswrapper[4959]: [+]process-running ok Oct 07 13:47:43 crc kubenswrapper[4959]: healthz check failed Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.826682 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nt7g" podUID="67543fb5-5742-4067-a953-7d94ac45e415" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.885007 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:43 crc kubenswrapper[4959]: E1007 13:47:43.886193 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:44.386173895 +0000 UTC m=+146.469578220 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.943354 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.945435 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.964066 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.986876 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:43 crc kubenswrapper[4959]: E1007 13:47:43.987094 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:44.487076372 +0000 UTC m=+146.570480697 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:43 crc kubenswrapper[4959]: I1007 13:47:43.987209 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:43 crc kubenswrapper[4959]: E1007 13:47:43.987562 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:44.487546085 +0000 UTC m=+146.570950490 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.087935 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:44 crc kubenswrapper[4959]: E1007 13:47:44.088249 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:44.588235056 +0000 UTC m=+146.671639381 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.189290 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:44 crc kubenswrapper[4959]: E1007 13:47:44.189635 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:44.689618786 +0000 UTC m=+146.773023111 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.261515 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ls9ng" event={"ID":"bd1bdacf-99ea-4afc-b804-2a87afb35cc7","Type":"ContainerStarted","Data":"e661c645abaa2375dc6ced21992e1416a398c2374197657f8c9e61bb4e57ae9c"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.261569 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ls9ng" event={"ID":"bd1bdacf-99ea-4afc-b804-2a87afb35cc7","Type":"ContainerStarted","Data":"1fa49e6bec55f18563fd92f1c72a420fe3de59be95b0fa7c742e12af98ba267f"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.261679 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ls9ng" Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.263470 4959 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-ls9ng container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" start-of-body= Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.263524 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ls9ng" podUID="bd1bdacf-99ea-4afc-b804-2a87afb35cc7" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.263814 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rfcqs" event={"ID":"48528215-02d5-48aa-8289-9defecf47166","Type":"ContainerStarted","Data":"3d740a0f2f281cc54466255f12ee05351d3c8d961d7a0a9f4b9c624d326ffd7f"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.283058 4959 generic.go:334] "Generic (PLEG): container finished" podID="e7167382-2267-4848-a423-e627253c38ed" containerID="4fc6a754c17020e11f3e6f467ea0a7f74c7a94bb6977c077295f3c81cca3d8d8" exitCode=0 Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.283195 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" event={"ID":"e7167382-2267-4848-a423-e627253c38ed","Type":"ContainerDied","Data":"4fc6a754c17020e11f3e6f467ea0a7f74c7a94bb6977c077295f3c81cca3d8d8"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.293063 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ls9ng" podStartSLOduration=120.293041861 podStartE2EDuration="2m0.293041861s" podCreationTimestamp="2025-10-07 13:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:44.292469185 +0000 UTC m=+146.375873510" watchObservedRunningTime="2025-10-07 13:47:44.293041861 +0000 UTC m=+146.376446186" Oct 07 13:47:44 crc kubenswrapper[4959]: E1007 13:47:44.293420 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:44.793395341 +0000 UTC m=+146.876799666 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.291594 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.296326 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:44 crc kubenswrapper[4959]: E1007 13:47:44.303172 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:44.80314663 +0000 UTC m=+146.886550955 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.307345 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-5qkt6" event={"ID":"09af6fe0-f3ab-4223-9b45-05509e575ac4","Type":"ContainerStarted","Data":"dcee4604556064bf005f2eba733d90f72a1c1c911c890ad63f648682cd56969c"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.307404 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-5qkt6" event={"ID":"09af6fe0-f3ab-4223-9b45-05509e575ac4","Type":"ContainerStarted","Data":"0374b49bdbcfaedcc1d531f8ac0cfca26b8806b1819fcd53f012061cde7c0676"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.333592 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-7bxtq" event={"ID":"8307e459-035b-4c33-bb8e-b5a698618036","Type":"ContainerStarted","Data":"45c20ad0625868f45d8fb0f8e266609362766eeda75d6053b3b87d1e4339c355"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.334417 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-7bxtq" Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.336783 4959 patch_prober.go:28] interesting pod/downloads-7954f5f757-7bxtq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.336829 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7bxtq" podUID="8307e459-035b-4c33-bb8e-b5a698618036" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.371190 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-7bxtq" podStartSLOduration=121.371172489 podStartE2EDuration="2m1.371172489s" podCreationTimestamp="2025-10-07 13:45:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:44.369647687 +0000 UTC m=+146.453052012" watchObservedRunningTime="2025-10-07 13:47:44.371172489 +0000 UTC m=+146.454576814" Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.377544 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lxbgb" event={"ID":"b748d317-5cb9-4fd3-a6bc-4a83d2f77d97","Type":"ContainerStarted","Data":"bf1898d2f45c1b56d2144392abb796f47ae532785a9335d5556e54f6fdd050fa"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.400897 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:44 crc kubenswrapper[4959]: E1007 13:47:44.403370 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:44.903343088 +0000 UTC m=+146.986747503 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.408326 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s95w5" event={"ID":"de5396bf-4d94-46be-87e1-7b587147a6a4","Type":"ContainerStarted","Data":"496b1731e9a7d92dcd45e946a7ec594c2dc109888be32ea9f541dceff4f8e34e"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.446151 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rxt22" event={"ID":"54540da0-9e56-4ce2-bf77-1d055296e4b1","Type":"ContainerStarted","Data":"1690805c4054b558d19e4396a04cad77b29318c70a8c4a20a3447d5d33806188"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.455971 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-fggsp" event={"ID":"07357227-cc26-43b8-89c7-4990980b2725","Type":"ContainerStarted","Data":"c86a7185b41354663bdca3973f52aef13ec36d1ea5f00f6a3504382d219476f6"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.465452 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lxbgb" podStartSLOduration=120.465436913 podStartE2EDuration="2m0.465436913s" podCreationTimestamp="2025-10-07 13:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:44.424009108 +0000 UTC m=+146.507413433" watchObservedRunningTime="2025-10-07 13:47:44.465436913 +0000 UTC m=+146.548841238" Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.482422 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-v5pq6" event={"ID":"468622a9-56e9-47f6-aa19-b9df629881b7","Type":"ContainerStarted","Data":"f621e0081bcdb4a7514fe608de532846a2b63d1c24c04b70fbb806281287d406"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.482473 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-v5pq6" event={"ID":"468622a9-56e9-47f6-aa19-b9df629881b7","Type":"ContainerStarted","Data":"0b28704fe6e4d2ad440e67af24b2289e8654c15f4141e9c86c25bfb11291cdfb"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.493737 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jj9rd" event={"ID":"c286e288-6157-4fff-bb4a-927e823d451a","Type":"ContainerStarted","Data":"40a01f765778efbc5da2a0e83860696f6be7c8ce0293ceb3c012b6b530b37300"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.493782 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jj9rd" event={"ID":"c286e288-6157-4fff-bb4a-927e823d451a","Type":"ContainerStarted","Data":"35ccde4c5cb9f4d5afa1b681460cc0ba9010d72ca648dfe36f7959c13169a387"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.504368 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:44 crc kubenswrapper[4959]: E1007 13:47:44.507187 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:45.007159665 +0000 UTC m=+147.090563980 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.515317 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zg2sm" event={"ID":"81d176a3-10ba-42f5-b716-d18a7dc1b5a8","Type":"ContainerStarted","Data":"3d67c3356a38d21e82a30d9b6d96b0237c231f0314c62b43888ace0262404698"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.518677 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rxt22" podStartSLOduration=120.518660803 podStartE2EDuration="2m0.518660803s" podCreationTimestamp="2025-10-07 13:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:44.463749076 +0000 UTC m=+146.547153401" watchObservedRunningTime="2025-10-07 13:47:44.518660803 +0000 UTC m=+146.602065128" Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.519344 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-v5pq6" podStartSLOduration=120.519340551 podStartE2EDuration="2m0.519340551s" podCreationTimestamp="2025-10-07 13:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:44.516236226 +0000 UTC m=+146.599640551" watchObservedRunningTime="2025-10-07 13:47:44.519340551 +0000 UTC m=+146.602744876" Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.528702 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gfnhv" event={"ID":"2843a931-4390-4dae-a736-911e2fc18bef","Type":"ContainerStarted","Data":"be9e07243718a06e93fcb50a89a45b3223c5d40827502d28d2fcad09ea10c7c9"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.529628 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gfnhv" Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.535202 4959 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-gfnhv container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.535256 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gfnhv" podUID="2843a931-4390-4dae-a736-911e2fc18bef" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.544973 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jj9rd" podStartSLOduration=120.544946178 podStartE2EDuration="2m0.544946178s" podCreationTimestamp="2025-10-07 13:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:44.543475198 +0000 UTC m=+146.626879523" watchObservedRunningTime="2025-10-07 13:47:44.544946178 +0000 UTC m=+146.628350493" Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.573237 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gfnhv" podStartSLOduration=120.573196019 podStartE2EDuration="2m0.573196019s" podCreationTimestamp="2025-10-07 13:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:44.563685256 +0000 UTC m=+146.647089581" watchObservedRunningTime="2025-10-07 13:47:44.573196019 +0000 UTC m=+146.656600344" Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.578258 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-cw2cl" event={"ID":"3218d84a-b760-4002-bf1f-79d02293ad93","Type":"ContainerStarted","Data":"173e6127ab48718c9b952c9b44aa0d19f9427018d8aac5b4b2ec45831c913bcc"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.578315 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-cw2cl" event={"ID":"3218d84a-b760-4002-bf1f-79d02293ad93","Type":"ContainerStarted","Data":"12ec548dd88e097aed8f5a84593f7eb55d88267c5fd3afb15de188d0fa2b68ad"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.587671 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xwlzm" event={"ID":"cb8de785-a298-414c-998c-7bd0585966e1","Type":"ContainerStarted","Data":"0127c5ab88c1d7a1993c76d818a43a226a5ee407b5f5aaa3ff686cc1d404aa63"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.587752 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xwlzm" event={"ID":"cb8de785-a298-414c-998c-7bd0585966e1","Type":"ContainerStarted","Data":"b5c6f3bda831fa6d2461f11a9f68d20440ee425b5e270cb2bd2bb004178e7670"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.588640 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-xwlzm" Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.594081 4959 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-xwlzm container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.594153 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-xwlzm" podUID="cb8de785-a298-414c-998c-7bd0585966e1" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.607751 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:44 crc kubenswrapper[4959]: E1007 13:47:44.608221 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:45.108199216 +0000 UTC m=+147.191603541 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.608944 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:44 crc kubenswrapper[4959]: E1007 13:47:44.610654 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:45.110636793 +0000 UTC m=+147.194041118 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.629554 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-xwlzm" podStartSLOduration=120.629534455 podStartE2EDuration="2m0.629534455s" podCreationTimestamp="2025-10-07 13:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:44.610053637 +0000 UTC m=+146.693457962" watchObservedRunningTime="2025-10-07 13:47:44.629534455 +0000 UTC m=+146.712938780" Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.632849 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-t8v4d" event={"ID":"62481657-1b38-4d8d-81d5-665bd2574467","Type":"ContainerStarted","Data":"5750c13e9608103a7796ad758099b38a7218e79ed41fefa73627a16b98cb7458"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.640571 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-7cvhs" event={"ID":"15866f74-4fd6-46df-b182-7cd58b5c1e0b","Type":"ContainerStarted","Data":"13a3b5aa986f6acbdc93c50d7273b6a433ae886696d436555c81d21dc12415a0"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.644442 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xkp55" event={"ID":"f25b8da6-8185-4faa-b800-3deef3a9b06b","Type":"ContainerStarted","Data":"e7cb1f560daff4f9d42781251579658b9ae2e86d3d3d1594d8c4aad63818ab06"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.644924 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xkp55" Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.646514 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z54z" event={"ID":"3b0dd795-8b21-43a2-9e50-b0cb668ac22c","Type":"ContainerStarted","Data":"ed018640708303c951fa594f40b94679360c94476e867b0f188a12917b7d0b21"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.647743 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z54z" Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.668851 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-7cvhs" podStartSLOduration=7.66883076 podStartE2EDuration="7.66883076s" podCreationTimestamp="2025-10-07 13:47:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:44.666911397 +0000 UTC m=+146.750315732" watchObservedRunningTime="2025-10-07 13:47:44.66883076 +0000 UTC m=+146.752235085" Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.676172 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-94zbw" event={"ID":"b816c481-6891-479c-99db-95f6e4abbc05","Type":"ContainerStarted","Data":"9304e3f7b5c12f11e26b9d57ac27ddb869bbfeb643f552539161618549a2af1a"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.676843 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-94zbw" Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.698764 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xkp55" podStartSLOduration=121.698749856 podStartE2EDuration="2m1.698749856s" podCreationTimestamp="2025-10-07 13:45:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:44.698122379 +0000 UTC m=+146.781526704" watchObservedRunningTime="2025-10-07 13:47:44.698749856 +0000 UTC m=+146.782154181" Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.701017 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2bh9c" event={"ID":"fe29debd-926a-4fd7-900c-b91a4d203436","Type":"ContainerStarted","Data":"6973323462e71c62b39485951474fda4501ac3f2b9c39cc1078fb36444cd1afc"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.709766 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:44 crc kubenswrapper[4959]: E1007 13:47:44.711005 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:45.210991594 +0000 UTC m=+147.294395919 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.737555 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z54z" podStartSLOduration=120.737536567 podStartE2EDuration="2m0.737536567s" podCreationTimestamp="2025-10-07 13:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:44.729488265 +0000 UTC m=+146.812892590" watchObservedRunningTime="2025-10-07 13:47:44.737536567 +0000 UTC m=+146.820940892" Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.738874 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-529ml" event={"ID":"932c7fc9-8796-41bc-982f-74dfd0d887be","Type":"ContainerStarted","Data":"b553f14c28622ae6e996b04dbb136dccaac79d78deccb4060d611dfc4794bff3"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.738907 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-529ml" event={"ID":"932c7fc9-8796-41bc-982f-74dfd0d887be","Type":"ContainerStarted","Data":"4d0f887ecfb99c803f0e401439f97b4ac5416a5e251dd0bb9974ccca05d7d8c5"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.738917 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-529ml" event={"ID":"932c7fc9-8796-41bc-982f-74dfd0d887be","Type":"ContainerStarted","Data":"d8327ce4daa817efbbf83b63397f2a1a94e4a3daf937d3a6eaeff697bec517cf"} Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.756189 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2bh9c" podStartSLOduration=121.756172742 podStartE2EDuration="2m1.756172742s" podCreationTimestamp="2025-10-07 13:45:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:44.753539719 +0000 UTC m=+146.836944044" watchObservedRunningTime="2025-10-07 13:47:44.756172742 +0000 UTC m=+146.839577067" Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.756258 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hsdts" Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.778296 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-94zbw" podStartSLOduration=120.778281203 podStartE2EDuration="2m0.778281203s" podCreationTimestamp="2025-10-07 13:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:44.776597076 +0000 UTC m=+146.860001401" watchObservedRunningTime="2025-10-07 13:47:44.778281203 +0000 UTC m=+146.861685528" Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.811433 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.811981 4959 patch_prober.go:28] interesting pod/router-default-5444994796-8nt7g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 13:47:44 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 07 13:47:44 crc kubenswrapper[4959]: [+]process-running ok Oct 07 13:47:44 crc kubenswrapper[4959]: healthz check failed Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.812030 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nt7g" podUID="67543fb5-5742-4067-a953-7d94ac45e415" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 13:47:44 crc kubenswrapper[4959]: E1007 13:47:44.813460 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:45.313443794 +0000 UTC m=+147.396848119 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.886014 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-529ml" podStartSLOduration=120.885979107 podStartE2EDuration="2m0.885979107s" podCreationTimestamp="2025-10-07 13:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:44.885409491 +0000 UTC m=+146.968813816" watchObservedRunningTime="2025-10-07 13:47:44.885979107 +0000 UTC m=+146.969383432" Oct 07 13:47:44 crc kubenswrapper[4959]: I1007 13:47:44.915528 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:44 crc kubenswrapper[4959]: E1007 13:47:44.915831 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:45.415817571 +0000 UTC m=+147.499221896 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.016716 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:45 crc kubenswrapper[4959]: E1007 13:47:45.017043 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:45.517031376 +0000 UTC m=+147.600435701 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.117970 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:45 crc kubenswrapper[4959]: E1007 13:47:45.118685 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:45.618669444 +0000 UTC m=+147.702073769 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.220049 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:45 crc kubenswrapper[4959]: E1007 13:47:45.220450 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:45.720424804 +0000 UTC m=+147.803829119 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.340503 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:45 crc kubenswrapper[4959]: E1007 13:47:45.340689 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:45.840658594 +0000 UTC m=+147.924062919 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.340897 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:45 crc kubenswrapper[4959]: E1007 13:47:45.341259 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:45.84124375 +0000 UTC m=+147.924648075 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.442330 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:45 crc kubenswrapper[4959]: E1007 13:47:45.442481 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:45.942461526 +0000 UTC m=+148.025865851 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.443058 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:45 crc kubenswrapper[4959]: E1007 13:47:45.443306 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:45.943298399 +0000 UTC m=+148.026702724 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.524188 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z54z" Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.544464 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:45 crc kubenswrapper[4959]: E1007 13:47:45.544666 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:46.044639287 +0000 UTC m=+148.128043612 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.544900 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:45 crc kubenswrapper[4959]: E1007 13:47:45.545354 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:46.045325106 +0000 UTC m=+148.128729431 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.645709 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:45 crc kubenswrapper[4959]: E1007 13:47:45.645893 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:46.145864793 +0000 UTC m=+148.229269118 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.646005 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:45 crc kubenswrapper[4959]: E1007 13:47:45.646353 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:46.146345816 +0000 UTC m=+148.229750141 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.745806 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-cw2cl" event={"ID":"3218d84a-b760-4002-bf1f-79d02293ad93","Type":"ContainerStarted","Data":"61e0d7baad08f5fc589516570ce6b14e085241ceee4f2b558b4e6ceb24cf030e"} Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.746442 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:45 crc kubenswrapper[4959]: E1007 13:47:45.746579 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:46.246558104 +0000 UTC m=+148.329962429 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.746742 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:45 crc kubenswrapper[4959]: E1007 13:47:45.746986 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:46.246978016 +0000 UTC m=+148.330382341 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.747610 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-fggsp" event={"ID":"07357227-cc26-43b8-89c7-4990980b2725","Type":"ContainerStarted","Data":"53cde9813d9ae49f320910fc992954ccc82efae6acad4ff75f83337e865bc099"} Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.748812 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rfcqs" event={"ID":"48528215-02d5-48aa-8289-9defecf47166","Type":"ContainerStarted","Data":"ba942cb9e67e9236377828831d5c183ac0b1ce02947ffbb67d6e16301e28ddcb"} Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.748956 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-rfcqs" Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.750457 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" event={"ID":"e7167382-2267-4848-a423-e627253c38ed","Type":"ContainerStarted","Data":"11338b62f29caebdf881d175bbbfbdfa6f6b02aebac938c31461430458b5b910"} Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.750483 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" event={"ID":"e7167382-2267-4848-a423-e627253c38ed","Type":"ContainerStarted","Data":"638cd19dbba630a8bbe922c45e418299b3e66a1dcac27d3bf4999f52b2b5f837"} Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.752207 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-5qkt6" event={"ID":"09af6fe0-f3ab-4223-9b45-05509e575ac4","Type":"ContainerStarted","Data":"9137001d872449f866911858da0d0255a9663f621ea903c35dd0fbdbb3479498"} Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.753573 4959 patch_prober.go:28] interesting pod/downloads-7954f5f757-7bxtq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.753620 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7bxtq" podUID="8307e459-035b-4c33-bb8e-b5a698618036" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.753849 4959 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-xwlzm container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.753881 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-xwlzm" podUID="cb8de785-a298-414c-998c-7bd0585966e1" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.757873 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gfnhv" Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.781367 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-cw2cl" podStartSLOduration=122.781336845 podStartE2EDuration="2m2.781336845s" podCreationTimestamp="2025-10-07 13:45:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:45.77283517 +0000 UTC m=+147.856239485" watchObservedRunningTime="2025-10-07 13:47:45.781336845 +0000 UTC m=+147.864741170" Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.803461 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ls9ng" Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.822488 4959 patch_prober.go:28] interesting pod/router-default-5444994796-8nt7g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 13:47:45 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 07 13:47:45 crc kubenswrapper[4959]: [+]process-running ok Oct 07 13:47:45 crc kubenswrapper[4959]: healthz check failed Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.822559 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nt7g" podUID="67543fb5-5742-4067-a953-7d94ac45e415" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.829411 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-rfcqs" podStartSLOduration=8.829388792 podStartE2EDuration="8.829388792s" podCreationTimestamp="2025-10-07 13:47:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:45.822729518 +0000 UTC m=+147.906133833" watchObservedRunningTime="2025-10-07 13:47:45.829388792 +0000 UTC m=+147.912793117" Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.850652 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:45 crc kubenswrapper[4959]: E1007 13:47:45.854166 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:46.354144965 +0000 UTC m=+148.437549290 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.863428 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" podStartSLOduration=122.863411551 podStartE2EDuration="2m2.863411551s" podCreationTimestamp="2025-10-07 13:45:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:45.863067002 +0000 UTC m=+147.946471317" watchObservedRunningTime="2025-10-07 13:47:45.863411551 +0000 UTC m=+147.946815866" Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.924804 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-5qkt6" podStartSLOduration=121.924787196 podStartE2EDuration="2m1.924787196s" podCreationTimestamp="2025-10-07 13:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:45.896541756 +0000 UTC m=+147.979946081" watchObservedRunningTime="2025-10-07 13:47:45.924787196 +0000 UTC m=+148.008191521" Oct 07 13:47:45 crc kubenswrapper[4959]: I1007 13:47:45.952731 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:45 crc kubenswrapper[4959]: E1007 13:47:45.953170 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:46.45315687 +0000 UTC m=+148.536561195 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.054497 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:46 crc kubenswrapper[4959]: E1007 13:47:46.054646 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:46.554626722 +0000 UTC m=+148.638031047 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.054762 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:46 crc kubenswrapper[4959]: E1007 13:47:46.055064 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:46.555055764 +0000 UTC m=+148.638460089 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.155865 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:46 crc kubenswrapper[4959]: E1007 13:47:46.155978 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:46.655962321 +0000 UTC m=+148.739366646 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.156472 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:46 crc kubenswrapper[4959]: E1007 13:47:46.156854 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:46.656837715 +0000 UTC m=+148.740242030 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.257607 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:46 crc kubenswrapper[4959]: E1007 13:47:46.257763 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:46.757737862 +0000 UTC m=+148.841142187 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.257892 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:46 crc kubenswrapper[4959]: E1007 13:47:46.258178 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:46.758165584 +0000 UTC m=+148.841569899 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.358773 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:46 crc kubenswrapper[4959]: E1007 13:47:46.359045 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:46.859001299 +0000 UTC m=+148.942405624 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.359344 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:46 crc kubenswrapper[4959]: E1007 13:47:46.359644 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:46.859629466 +0000 UTC m=+148.943033791 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.460388 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:46 crc kubenswrapper[4959]: E1007 13:47:46.460603 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:46.960571494 +0000 UTC m=+149.043975819 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.460666 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:46 crc kubenswrapper[4959]: E1007 13:47:46.461009 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:46.960998836 +0000 UTC m=+149.044403161 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.487635 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5r424"] Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.488625 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5r424" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.491076 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.522577 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5r424"] Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.561836 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:46 crc kubenswrapper[4959]: E1007 13:47:46.562046 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:47.062013416 +0000 UTC m=+149.145417741 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.562235 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.562276 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhdvl\" (UniqueName: \"kubernetes.io/projected/4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac-kube-api-access-qhdvl\") pod \"certified-operators-5r424\" (UID: \"4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac\") " pod="openshift-marketplace/certified-operators-5r424" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.562370 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac-catalog-content\") pod \"certified-operators-5r424\" (UID: \"4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac\") " pod="openshift-marketplace/certified-operators-5r424" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.562401 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac-utilities\") pod \"certified-operators-5r424\" (UID: \"4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac\") " pod="openshift-marketplace/certified-operators-5r424" Oct 07 13:47:46 crc kubenswrapper[4959]: E1007 13:47:46.562581 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:47.062573871 +0000 UTC m=+149.145978196 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.574933 4959 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.659905 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t7mrz"] Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.660978 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t7mrz" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.662896 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.663131 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac-catalog-content\") pod \"certified-operators-5r424\" (UID: \"4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac\") " pod="openshift-marketplace/certified-operators-5r424" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.663152 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac-utilities\") pod \"certified-operators-5r424\" (UID: \"4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac\") " pod="openshift-marketplace/certified-operators-5r424" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.663198 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhdvl\" (UniqueName: \"kubernetes.io/projected/4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac-kube-api-access-qhdvl\") pod \"certified-operators-5r424\" (UID: \"4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac\") " pod="openshift-marketplace/certified-operators-5r424" Oct 07 13:47:46 crc kubenswrapper[4959]: E1007 13:47:46.663434 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-07 13:47:47.163373235 +0000 UTC m=+149.246777560 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.663638 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac-catalog-content\") pod \"certified-operators-5r424\" (UID: \"4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac\") " pod="openshift-marketplace/certified-operators-5r424" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.663706 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac-utilities\") pod \"certified-operators-5r424\" (UID: \"4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac\") " pod="openshift-marketplace/certified-operators-5r424" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.663913 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.664797 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t7mrz"] Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.698602 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhdvl\" (UniqueName: \"kubernetes.io/projected/4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac-kube-api-access-qhdvl\") pod \"certified-operators-5r424\" (UID: \"4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac\") " pod="openshift-marketplace/certified-operators-5r424" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.764047 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebb5441f-839a-408b-bd51-3460006c3d83-catalog-content\") pod \"community-operators-t7mrz\" (UID: \"ebb5441f-839a-408b-bd51-3460006c3d83\") " pod="openshift-marketplace/community-operators-t7mrz" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.764089 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.764190 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.764228 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vvdz\" (UniqueName: \"kubernetes.io/projected/ebb5441f-839a-408b-bd51-3460006c3d83-kube-api-access-7vvdz\") pod \"community-operators-t7mrz\" (UID: \"ebb5441f-839a-408b-bd51-3460006c3d83\") " pod="openshift-marketplace/community-operators-t7mrz" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.764254 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.764282 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.764327 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.764351 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebb5441f-839a-408b-bd51-3460006c3d83-utilities\") pod \"community-operators-t7mrz\" (UID: \"ebb5441f-839a-408b-bd51-3460006c3d83\") " pod="openshift-marketplace/community-operators-t7mrz" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.778787 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.778882 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:47:46 crc kubenswrapper[4959]: E1007 13:47:46.779545 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-07 13:47:47.279529383 +0000 UTC m=+149.362933708 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jhgsq" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.787916 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.801616 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5r424" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.804244 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.805265 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-fggsp" event={"ID":"07357227-cc26-43b8-89c7-4990980b2725","Type":"ContainerStarted","Data":"aa62828f0b4b637519074dac094b74f6e5a06400c574f55735c39ae5b6cda7e7"} Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.805305 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-fggsp" event={"ID":"07357227-cc26-43b8-89c7-4990980b2725","Type":"ContainerStarted","Data":"ae8b09604b702bc3b4e22b84f2064239d1198d3de3059bd631e12f3306b4439f"} Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.809209 4959 patch_prober.go:28] interesting pod/downloads-7954f5f757-7bxtq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.809264 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7bxtq" podUID="8307e459-035b-4c33-bb8e-b5a698618036" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.819327 4959 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-07T13:47:46.574961103Z","Handler":null,"Name":""} Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.832132 4959 patch_prober.go:28] interesting pod/router-default-5444994796-8nt7g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 13:47:46 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 07 13:47:46 crc kubenswrapper[4959]: [+]process-running ok Oct 07 13:47:46 crc kubenswrapper[4959]: healthz check failed Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.832181 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nt7g" podUID="67543fb5-5742-4067-a953-7d94ac45e415" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.857984 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9bkvf"] Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.860228 4959 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.860263 4959 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.865294 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9bkvf" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.879250 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9bkvf"] Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.879660 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.879965 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vvdz\" (UniqueName: \"kubernetes.io/projected/ebb5441f-839a-408b-bd51-3460006c3d83-kube-api-access-7vvdz\") pod \"community-operators-t7mrz\" (UID: \"ebb5441f-839a-408b-bd51-3460006c3d83\") " pod="openshift-marketplace/community-operators-t7mrz" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.880132 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebb5441f-839a-408b-bd51-3460006c3d83-utilities\") pod \"community-operators-t7mrz\" (UID: \"ebb5441f-839a-408b-bd51-3460006c3d83\") " pod="openshift-marketplace/community-operators-t7mrz" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.880242 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebb5441f-839a-408b-bd51-3460006c3d83-catalog-content\") pod \"community-operators-t7mrz\" (UID: \"ebb5441f-839a-408b-bd51-3460006c3d83\") " pod="openshift-marketplace/community-operators-t7mrz" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.880637 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebb5441f-839a-408b-bd51-3460006c3d83-catalog-content\") pod \"community-operators-t7mrz\" (UID: \"ebb5441f-839a-408b-bd51-3460006c3d83\") " pod="openshift-marketplace/community-operators-t7mrz" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.880957 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebb5441f-839a-408b-bd51-3460006c3d83-utilities\") pod \"community-operators-t7mrz\" (UID: \"ebb5441f-839a-408b-bd51-3460006c3d83\") " pod="openshift-marketplace/community-operators-t7mrz" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.897261 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.912825 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vvdz\" (UniqueName: \"kubernetes.io/projected/ebb5441f-839a-408b-bd51-3460006c3d83-kube-api-access-7vvdz\") pod \"community-operators-t7mrz\" (UID: \"ebb5441f-839a-408b-bd51-3460006c3d83\") " pod="openshift-marketplace/community-operators-t7mrz" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.982869 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t7mrz" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.983029 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw9ct\" (UniqueName: \"kubernetes.io/projected/f0e27f2e-7704-4082-91e5-571830f218fc-kube-api-access-rw9ct\") pod \"certified-operators-9bkvf\" (UID: \"f0e27f2e-7704-4082-91e5-571830f218fc\") " pod="openshift-marketplace/certified-operators-9bkvf" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.983115 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.983164 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0e27f2e-7704-4082-91e5-571830f218fc-utilities\") pod \"certified-operators-9bkvf\" (UID: \"f0e27f2e-7704-4082-91e5-571830f218fc\") " pod="openshift-marketplace/certified-operators-9bkvf" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.983181 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0e27f2e-7704-4082-91e5-571830f218fc-catalog-content\") pod \"certified-operators-9bkvf\" (UID: \"f0e27f2e-7704-4082-91e5-571830f218fc\") " pod="openshift-marketplace/certified-operators-9bkvf" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.983952 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 07 13:47:46 crc kubenswrapper[4959]: I1007 13:47:46.998026 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.002372 4959 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.002414 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.055258 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kkxdr"] Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.056687 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kkxdr" Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.074453 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kkxdr"] Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.079213 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.081530 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jhgsq\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.083796 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0e27f2e-7704-4082-91e5-571830f218fc-utilities\") pod \"certified-operators-9bkvf\" (UID: \"f0e27f2e-7704-4082-91e5-571830f218fc\") " pod="openshift-marketplace/certified-operators-9bkvf" Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.083823 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0e27f2e-7704-4082-91e5-571830f218fc-catalog-content\") pod \"certified-operators-9bkvf\" (UID: \"f0e27f2e-7704-4082-91e5-571830f218fc\") " pod="openshift-marketplace/certified-operators-9bkvf" Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.083862 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw9ct\" (UniqueName: \"kubernetes.io/projected/f0e27f2e-7704-4082-91e5-571830f218fc-kube-api-access-rw9ct\") pod \"certified-operators-9bkvf\" (UID: \"f0e27f2e-7704-4082-91e5-571830f218fc\") " pod="openshift-marketplace/certified-operators-9bkvf" Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.084856 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0e27f2e-7704-4082-91e5-571830f218fc-utilities\") pod \"certified-operators-9bkvf\" (UID: \"f0e27f2e-7704-4082-91e5-571830f218fc\") " pod="openshift-marketplace/certified-operators-9bkvf" Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.085061 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0e27f2e-7704-4082-91e5-571830f218fc-catalog-content\") pod \"certified-operators-9bkvf\" (UID: \"f0e27f2e-7704-4082-91e5-571830f218fc\") " pod="openshift-marketplace/certified-operators-9bkvf" Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.104729 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw9ct\" (UniqueName: \"kubernetes.io/projected/f0e27f2e-7704-4082-91e5-571830f218fc-kube-api-access-rw9ct\") pod \"certified-operators-9bkvf\" (UID: \"f0e27f2e-7704-4082-91e5-571830f218fc\") " pod="openshift-marketplace/certified-operators-9bkvf" Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.110330 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5r424"] Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.181693 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9bkvf" Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.184901 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grzcm\" (UniqueName: \"kubernetes.io/projected/a5295901-596b-447c-8d7a-bac77b140153-kube-api-access-grzcm\") pod \"community-operators-kkxdr\" (UID: \"a5295901-596b-447c-8d7a-bac77b140153\") " pod="openshift-marketplace/community-operators-kkxdr" Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.184937 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5295901-596b-447c-8d7a-bac77b140153-catalog-content\") pod \"community-operators-kkxdr\" (UID: \"a5295901-596b-447c-8d7a-bac77b140153\") " pod="openshift-marketplace/community-operators-kkxdr" Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.184957 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5295901-596b-447c-8d7a-bac77b140153-utilities\") pod \"community-operators-kkxdr\" (UID: \"a5295901-596b-447c-8d7a-bac77b140153\") " pod="openshift-marketplace/community-operators-kkxdr" Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.286094 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grzcm\" (UniqueName: \"kubernetes.io/projected/a5295901-596b-447c-8d7a-bac77b140153-kube-api-access-grzcm\") pod \"community-operators-kkxdr\" (UID: \"a5295901-596b-447c-8d7a-bac77b140153\") " pod="openshift-marketplace/community-operators-kkxdr" Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.286486 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5295901-596b-447c-8d7a-bac77b140153-catalog-content\") pod \"community-operators-kkxdr\" (UID: \"a5295901-596b-447c-8d7a-bac77b140153\") " pod="openshift-marketplace/community-operators-kkxdr" Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.286507 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5295901-596b-447c-8d7a-bac77b140153-utilities\") pod \"community-operators-kkxdr\" (UID: \"a5295901-596b-447c-8d7a-bac77b140153\") " pod="openshift-marketplace/community-operators-kkxdr" Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.287004 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5295901-596b-447c-8d7a-bac77b140153-utilities\") pod \"community-operators-kkxdr\" (UID: \"a5295901-596b-447c-8d7a-bac77b140153\") " pod="openshift-marketplace/community-operators-kkxdr" Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.287128 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5295901-596b-447c-8d7a-bac77b140153-catalog-content\") pod \"community-operators-kkxdr\" (UID: \"a5295901-596b-447c-8d7a-bac77b140153\") " pod="openshift-marketplace/community-operators-kkxdr" Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.309638 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grzcm\" (UniqueName: \"kubernetes.io/projected/a5295901-596b-447c-8d7a-bac77b140153-kube-api-access-grzcm\") pod \"community-operators-kkxdr\" (UID: \"a5295901-596b-447c-8d7a-bac77b140153\") " pod="openshift-marketplace/community-operators-kkxdr" Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.312874 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.389778 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kkxdr" Oct 07 13:47:47 crc kubenswrapper[4959]: W1007 13:47:47.476300 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-e9293a8dc5fea0af968f58d12c60425efbb67a8dee7bee9152b738d36b5b2a07 WatchSource:0}: Error finding container e9293a8dc5fea0af968f58d12c60425efbb67a8dee7bee9152b738d36b5b2a07: Status 404 returned error can't find the container with id e9293a8dc5fea0af968f58d12c60425efbb67a8dee7bee9152b738d36b5b2a07 Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.729992 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t7mrz"] Oct 07 13:47:47 crc kubenswrapper[4959]: W1007 13:47:47.749128 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podebb5441f_839a_408b_bd51_3460006c3d83.slice/crio-07fad97d5e8a0ee517b2c0b72e6dbe1b1f2a868e708e11455ab0cf1da003c6a9 WatchSource:0}: Error finding container 07fad97d5e8a0ee517b2c0b72e6dbe1b1f2a868e708e11455ab0cf1da003c6a9: Status 404 returned error can't find the container with id 07fad97d5e8a0ee517b2c0b72e6dbe1b1f2a868e708e11455ab0cf1da003c6a9 Oct 07 13:47:47 crc kubenswrapper[4959]: W1007 13:47:47.753289 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-d364fad8abbabf9f7336d18a9a824aa01f803fa54e24e62d8ab727e94e107c4d WatchSource:0}: Error finding container d364fad8abbabf9f7336d18a9a824aa01f803fa54e24e62d8ab727e94e107c4d: Status 404 returned error can't find the container with id d364fad8abbabf9f7336d18a9a824aa01f803fa54e24e62d8ab727e94e107c4d Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.800834 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9bkvf"] Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.811230 4959 patch_prober.go:28] interesting pod/router-default-5444994796-8nt7g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 13:47:47 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 07 13:47:47 crc kubenswrapper[4959]: [+]process-running ok Oct 07 13:47:47 crc kubenswrapper[4959]: healthz check failed Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.811281 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nt7g" podUID="67543fb5-5742-4067-a953-7d94ac45e415" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.819292 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9bkvf" event={"ID":"f0e27f2e-7704-4082-91e5-571830f218fc","Type":"ContainerStarted","Data":"7d08893e260f23e4dd96c0b2263c6acbc2f4f9a7be8b06b413570bf468519447"} Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.836268 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-fggsp" event={"ID":"07357227-cc26-43b8-89c7-4990980b2725","Type":"ContainerStarted","Data":"4c9f00bf33de201f5f4caf2d05e216dcccb311822e7a6228339fbecf8d5c4c4a"} Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.852547 4959 generic.go:334] "Generic (PLEG): container finished" podID="d0a596b0-4162-4b10-8880-3ea0421b8c74" containerID="a7017d01f13370140f49dfc5cb484478c43956cb32641304263fe6937e7d534b" exitCode=0 Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.852654 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330745-wbwkx" event={"ID":"d0a596b0-4162-4b10-8880-3ea0421b8c74","Type":"ContainerDied","Data":"a7017d01f13370140f49dfc5cb484478c43956cb32641304263fe6937e7d534b"} Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.893771 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"b428396626a42fdfbc15b7751f8bba58eba4d15c2f020b11b09ee1681ec25e0a"} Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.893820 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"cdea3be319d8d9ab0aae2e015941e000417937603e429853f722b2c0f9b1704e"} Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.895951 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"d364fad8abbabf9f7336d18a9a824aa01f803fa54e24e62d8ab727e94e107c4d"} Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.901373 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.902014 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.902835 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7mrz" event={"ID":"ebb5441f-839a-408b-bd51-3460006c3d83","Type":"ContainerStarted","Data":"07fad97d5e8a0ee517b2c0b72e6dbe1b1f2a868e708e11455ab0cf1da003c6a9"} Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.905566 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.905570 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.905969 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"b4511339208b7994529ac621f5450eb29effea7661ec0e278cb3515c767c9a08"} Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.905998 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"e9293a8dc5fea0af968f58d12c60425efbb67a8dee7bee9152b738d36b5b2a07"} Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.906746 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-fggsp" podStartSLOduration=10.906727774 podStartE2EDuration="10.906727774s" podCreationTimestamp="2025-10-07 13:47:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:47.902684662 +0000 UTC m=+149.986088987" watchObservedRunningTime="2025-10-07 13:47:47.906727774 +0000 UTC m=+149.990132099" Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.910986 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.922581 4959 generic.go:334] "Generic (PLEG): container finished" podID="4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac" containerID="cd437d6fec756cb435f8df059a24775e70036e577fc980e0cb1fb9e38fbbeea7" exitCode=0 Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.923844 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5r424" event={"ID":"4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac","Type":"ContainerDied","Data":"cd437d6fec756cb435f8df059a24775e70036e577fc980e0cb1fb9e38fbbeea7"} Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.923867 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5r424" event={"ID":"4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac","Type":"ContainerStarted","Data":"a6359874f8f2aa856f62bb09b77413cf993b7de19e61e8b2144bdc2ea74c4a08"} Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.924407 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jhgsq"] Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.933239 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kkxdr"] Oct 07 13:47:47 crc kubenswrapper[4959]: I1007 13:47:47.935624 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.000560 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/601096a4-0f0e-4c06-bce3-bdccdd1e370b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"601096a4-0f0e-4c06-bce3-bdccdd1e370b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.000684 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/601096a4-0f0e-4c06-bce3-bdccdd1e370b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"601096a4-0f0e-4c06-bce3-bdccdd1e370b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.101610 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/601096a4-0f0e-4c06-bce3-bdccdd1e370b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"601096a4-0f0e-4c06-bce3-bdccdd1e370b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.101692 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/601096a4-0f0e-4c06-bce3-bdccdd1e370b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"601096a4-0f0e-4c06-bce3-bdccdd1e370b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.101789 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/601096a4-0f0e-4c06-bce3-bdccdd1e370b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"601096a4-0f0e-4c06-bce3-bdccdd1e370b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.119405 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/601096a4-0f0e-4c06-bce3-bdccdd1e370b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"601096a4-0f0e-4c06-bce3-bdccdd1e370b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.248248 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.451624 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c94wx"] Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.452957 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c94wx" Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.458925 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c94wx"] Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.460705 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.500883 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 07 13:47:48 crc kubenswrapper[4959]: W1007 13:47:48.520375 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod601096a4_0f0e_4c06_bce3_bdccdd1e370b.slice/crio-56657ce5bc739a96605641045ceb210b12af55764bac692c91f2e835611cfef7 WatchSource:0}: Error finding container 56657ce5bc739a96605641045ceb210b12af55764bac692c91f2e835611cfef7: Status 404 returned error can't find the container with id 56657ce5bc739a96605641045ceb210b12af55764bac692c91f2e835611cfef7 Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.608484 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44b4f9d4-243a-46a8-ab8a-4fb5ae11816b-utilities\") pod \"redhat-marketplace-c94wx\" (UID: \"44b4f9d4-243a-46a8-ab8a-4fb5ae11816b\") " pod="openshift-marketplace/redhat-marketplace-c94wx" Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.608726 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44b4f9d4-243a-46a8-ab8a-4fb5ae11816b-catalog-content\") pod \"redhat-marketplace-c94wx\" (UID: \"44b4f9d4-243a-46a8-ab8a-4fb5ae11816b\") " pod="openshift-marketplace/redhat-marketplace-c94wx" Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.608784 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-925d4\" (UniqueName: \"kubernetes.io/projected/44b4f9d4-243a-46a8-ab8a-4fb5ae11816b-kube-api-access-925d4\") pod \"redhat-marketplace-c94wx\" (UID: \"44b4f9d4-243a-46a8-ab8a-4fb5ae11816b\") " pod="openshift-marketplace/redhat-marketplace-c94wx" Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.663361 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.722814 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-925d4\" (UniqueName: \"kubernetes.io/projected/44b4f9d4-243a-46a8-ab8a-4fb5ae11816b-kube-api-access-925d4\") pod \"redhat-marketplace-c94wx\" (UID: \"44b4f9d4-243a-46a8-ab8a-4fb5ae11816b\") " pod="openshift-marketplace/redhat-marketplace-c94wx" Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.722869 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44b4f9d4-243a-46a8-ab8a-4fb5ae11816b-utilities\") pod \"redhat-marketplace-c94wx\" (UID: \"44b4f9d4-243a-46a8-ab8a-4fb5ae11816b\") " pod="openshift-marketplace/redhat-marketplace-c94wx" Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.722951 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44b4f9d4-243a-46a8-ab8a-4fb5ae11816b-catalog-content\") pod \"redhat-marketplace-c94wx\" (UID: \"44b4f9d4-243a-46a8-ab8a-4fb5ae11816b\") " pod="openshift-marketplace/redhat-marketplace-c94wx" Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.724559 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44b4f9d4-243a-46a8-ab8a-4fb5ae11816b-catalog-content\") pod \"redhat-marketplace-c94wx\" (UID: \"44b4f9d4-243a-46a8-ab8a-4fb5ae11816b\") " pod="openshift-marketplace/redhat-marketplace-c94wx" Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.724777 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44b4f9d4-243a-46a8-ab8a-4fb5ae11816b-utilities\") pod \"redhat-marketplace-c94wx\" (UID: \"44b4f9d4-243a-46a8-ab8a-4fb5ae11816b\") " pod="openshift-marketplace/redhat-marketplace-c94wx" Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.743192 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-925d4\" (UniqueName: \"kubernetes.io/projected/44b4f9d4-243a-46a8-ab8a-4fb5ae11816b-kube-api-access-925d4\") pod \"redhat-marketplace-c94wx\" (UID: \"44b4f9d4-243a-46a8-ab8a-4fb5ae11816b\") " pod="openshift-marketplace/redhat-marketplace-c94wx" Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.771180 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c94wx" Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.810460 4959 patch_prober.go:28] interesting pod/router-default-5444994796-8nt7g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 13:47:48 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 07 13:47:48 crc kubenswrapper[4959]: [+]process-running ok Oct 07 13:47:48 crc kubenswrapper[4959]: healthz check failed Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.810843 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nt7g" podUID="67543fb5-5742-4067-a953-7d94ac45e415" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.847706 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5brqg"] Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.849152 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5brqg" Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.855738 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5brqg"] Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.925729 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11da31c7-46e5-4d9c-b11d-b67fae805f5d-utilities\") pod \"redhat-marketplace-5brqg\" (UID: \"11da31c7-46e5-4d9c-b11d-b67fae805f5d\") " pod="openshift-marketplace/redhat-marketplace-5brqg" Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.925789 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11da31c7-46e5-4d9c-b11d-b67fae805f5d-catalog-content\") pod \"redhat-marketplace-5brqg\" (UID: \"11da31c7-46e5-4d9c-b11d-b67fae805f5d\") " pod="openshift-marketplace/redhat-marketplace-5brqg" Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.925825 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk75g\" (UniqueName: \"kubernetes.io/projected/11da31c7-46e5-4d9c-b11d-b67fae805f5d-kube-api-access-tk75g\") pod \"redhat-marketplace-5brqg\" (UID: \"11da31c7-46e5-4d9c-b11d-b67fae805f5d\") " pod="openshift-marketplace/redhat-marketplace-5brqg" Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.955031 4959 generic.go:334] "Generic (PLEG): container finished" podID="a5295901-596b-447c-8d7a-bac77b140153" containerID="d310dd3b306a87bc16520f59be6687984d547c6d3eadc02cd9d9487dab392380" exitCode=0 Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.955129 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkxdr" event={"ID":"a5295901-596b-447c-8d7a-bac77b140153","Type":"ContainerDied","Data":"d310dd3b306a87bc16520f59be6687984d547c6d3eadc02cd9d9487dab392380"} Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.955166 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkxdr" event={"ID":"a5295901-596b-447c-8d7a-bac77b140153","Type":"ContainerStarted","Data":"0f2e047cc089b406d7dd7577703ae16a1ffe2a89e2b898e814114fe5bc0601f9"} Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.969587 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"65e8c19476b9a25aed6699af9ae0bb7dcaa420ec690dded17f6b8985b1ab3b3e"} Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.970477 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.972275 4959 generic.go:334] "Generic (PLEG): container finished" podID="ebb5441f-839a-408b-bd51-3460006c3d83" containerID="1f490ddca97ee825004567944d6c8291950ca819b496df51ca9bcb47e877e88d" exitCode=0 Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.972323 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7mrz" event={"ID":"ebb5441f-839a-408b-bd51-3460006c3d83","Type":"ContainerDied","Data":"1f490ddca97ee825004567944d6c8291950ca819b496df51ca9bcb47e877e88d"} Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.975392 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"601096a4-0f0e-4c06-bce3-bdccdd1e370b","Type":"ContainerStarted","Data":"56657ce5bc739a96605641045ceb210b12af55764bac692c91f2e835611cfef7"} Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.976864 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" event={"ID":"24ccfcc4-25a6-4264-86b1-6ba0097b2614","Type":"ContainerStarted","Data":"7e3b9068afcbcca93532dab4dc19a188ed50a6def356908e91711b5782b6ea4f"} Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.976894 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" event={"ID":"24ccfcc4-25a6-4264-86b1-6ba0097b2614","Type":"ContainerStarted","Data":"cdd43bc7abf0927692c795123bba03ab97784ecebbc50924418df50031e61079"} Oct 07 13:47:48 crc kubenswrapper[4959]: I1007 13:47:48.977790 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.000701 4959 generic.go:334] "Generic (PLEG): container finished" podID="f0e27f2e-7704-4082-91e5-571830f218fc" containerID="df3f22c056ab913f07fe26d3d08be8852e1249beabfafa9ee7f36c96334a4ad9" exitCode=0 Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.001821 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9bkvf" event={"ID":"f0e27f2e-7704-4082-91e5-571830f218fc","Type":"ContainerDied","Data":"df3f22c056ab913f07fe26d3d08be8852e1249beabfafa9ee7f36c96334a4ad9"} Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.016650 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" podStartSLOduration=125.016637648 podStartE2EDuration="2m5.016637648s" podCreationTimestamp="2025-10-07 13:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:47:49.01455128 +0000 UTC m=+151.097955605" watchObservedRunningTime="2025-10-07 13:47:49.016637648 +0000 UTC m=+151.100041973" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.030045 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11da31c7-46e5-4d9c-b11d-b67fae805f5d-utilities\") pod \"redhat-marketplace-5brqg\" (UID: \"11da31c7-46e5-4d9c-b11d-b67fae805f5d\") " pod="openshift-marketplace/redhat-marketplace-5brqg" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.030118 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11da31c7-46e5-4d9c-b11d-b67fae805f5d-catalog-content\") pod \"redhat-marketplace-5brqg\" (UID: \"11da31c7-46e5-4d9c-b11d-b67fae805f5d\") " pod="openshift-marketplace/redhat-marketplace-5brqg" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.030150 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk75g\" (UniqueName: \"kubernetes.io/projected/11da31c7-46e5-4d9c-b11d-b67fae805f5d-kube-api-access-tk75g\") pod \"redhat-marketplace-5brqg\" (UID: \"11da31c7-46e5-4d9c-b11d-b67fae805f5d\") " pod="openshift-marketplace/redhat-marketplace-5brqg" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.031228 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11da31c7-46e5-4d9c-b11d-b67fae805f5d-utilities\") pod \"redhat-marketplace-5brqg\" (UID: \"11da31c7-46e5-4d9c-b11d-b67fae805f5d\") " pod="openshift-marketplace/redhat-marketplace-5brqg" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.031504 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11da31c7-46e5-4d9c-b11d-b67fae805f5d-catalog-content\") pod \"redhat-marketplace-5brqg\" (UID: \"11da31c7-46e5-4d9c-b11d-b67fae805f5d\") " pod="openshift-marketplace/redhat-marketplace-5brqg" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.067188 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk75g\" (UniqueName: \"kubernetes.io/projected/11da31c7-46e5-4d9c-b11d-b67fae805f5d-kube-api-access-tk75g\") pod \"redhat-marketplace-5brqg\" (UID: \"11da31c7-46e5-4d9c-b11d-b67fae805f5d\") " pod="openshift-marketplace/redhat-marketplace-5brqg" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.127597 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c94wx"] Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.137004 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xkp55" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.165047 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5brqg" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.308986 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330745-wbwkx" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.435872 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9d4h\" (UniqueName: \"kubernetes.io/projected/d0a596b0-4162-4b10-8880-3ea0421b8c74-kube-api-access-j9d4h\") pod \"d0a596b0-4162-4b10-8880-3ea0421b8c74\" (UID: \"d0a596b0-4162-4b10-8880-3ea0421b8c74\") " Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.435955 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0a596b0-4162-4b10-8880-3ea0421b8c74-config-volume\") pod \"d0a596b0-4162-4b10-8880-3ea0421b8c74\" (UID: \"d0a596b0-4162-4b10-8880-3ea0421b8c74\") " Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.435992 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0a596b0-4162-4b10-8880-3ea0421b8c74-secret-volume\") pod \"d0a596b0-4162-4b10-8880-3ea0421b8c74\" (UID: \"d0a596b0-4162-4b10-8880-3ea0421b8c74\") " Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.437878 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0a596b0-4162-4b10-8880-3ea0421b8c74-config-volume" (OuterVolumeSpecName: "config-volume") pod "d0a596b0-4162-4b10-8880-3ea0421b8c74" (UID: "d0a596b0-4162-4b10-8880-3ea0421b8c74"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.446461 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0a596b0-4162-4b10-8880-3ea0421b8c74-kube-api-access-j9d4h" (OuterVolumeSpecName: "kube-api-access-j9d4h") pod "d0a596b0-4162-4b10-8880-3ea0421b8c74" (UID: "d0a596b0-4162-4b10-8880-3ea0421b8c74"). InnerVolumeSpecName "kube-api-access-j9d4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.451255 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5brqg"] Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.460266 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0a596b0-4162-4b10-8880-3ea0421b8c74-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d0a596b0-4162-4b10-8880-3ea0421b8c74" (UID: "d0a596b0-4162-4b10-8880-3ea0421b8c74"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:47:49 crc kubenswrapper[4959]: W1007 13:47:49.486596 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod11da31c7_46e5_4d9c_b11d_b67fae805f5d.slice/crio-d36ad2c834f36017d2e03645259f49c7465c493686d70a19d0bf9060e7a03ce1 WatchSource:0}: Error finding container d36ad2c834f36017d2e03645259f49c7465c493686d70a19d0bf9060e7a03ce1: Status 404 returned error can't find the container with id d36ad2c834f36017d2e03645259f49c7465c493686d70a19d0bf9060e7a03ce1 Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.537718 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9d4h\" (UniqueName: \"kubernetes.io/projected/d0a596b0-4162-4b10-8880-3ea0421b8c74-kube-api-access-j9d4h\") on node \"crc\" DevicePath \"\"" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.537755 4959 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0a596b0-4162-4b10-8880-3ea0421b8c74-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.537765 4959 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0a596b0-4162-4b10-8880-3ea0421b8c74-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.679840 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-d6vp9"] Oct 07 13:47:49 crc kubenswrapper[4959]: E1007 13:47:49.680190 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0a596b0-4162-4b10-8880-3ea0421b8c74" containerName="collect-profiles" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.680214 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0a596b0-4162-4b10-8880-3ea0421b8c74" containerName="collect-profiles" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.680338 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0a596b0-4162-4b10-8880-3ea0421b8c74" containerName="collect-profiles" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.686550 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d6vp9"] Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.686696 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d6vp9" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.706830 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.741884 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw8q2\" (UniqueName: \"kubernetes.io/projected/339283f2-dc57-4316-a334-6d79fbf2be9b-kube-api-access-zw8q2\") pod \"redhat-operators-d6vp9\" (UID: \"339283f2-dc57-4316-a334-6d79fbf2be9b\") " pod="openshift-marketplace/redhat-operators-d6vp9" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.741992 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/339283f2-dc57-4316-a334-6d79fbf2be9b-catalog-content\") pod \"redhat-operators-d6vp9\" (UID: \"339283f2-dc57-4316-a334-6d79fbf2be9b\") " pod="openshift-marketplace/redhat-operators-d6vp9" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.742081 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/339283f2-dc57-4316-a334-6d79fbf2be9b-utilities\") pod \"redhat-operators-d6vp9\" (UID: \"339283f2-dc57-4316-a334-6d79fbf2be9b\") " pod="openshift-marketplace/redhat-operators-d6vp9" Oct 07 13:47:49 crc kubenswrapper[4959]: E1007 13:47:49.759381 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-pod601096a4_0f0e_4c06_bce3_bdccdd1e370b.slice/crio-conmon-642608cc7d5c158344869ed239420aabbd43174bc508a958b0f238a8a518904b.scope\": RecentStats: unable to find data in memory cache]" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.797758 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-pcvlp" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.797813 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-pcvlp" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.799431 4959 patch_prober.go:28] interesting pod/console-f9d7485db-pcvlp container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.19:8443/health\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.799513 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-pcvlp" podUID="4fca2a0e-545f-43b0-b4ec-e5f15babcd71" containerName="console" probeResult="failure" output="Get \"https://10.217.0.19:8443/health\": dial tcp 10.217.0.19:8443: connect: connection refused" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.809219 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-8nt7g" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.814500 4959 patch_prober.go:28] interesting pod/router-default-5444994796-8nt7g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 13:47:49 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 07 13:47:49 crc kubenswrapper[4959]: [+]process-running ok Oct 07 13:47:49 crc kubenswrapper[4959]: healthz check failed Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.814591 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nt7g" podUID="67543fb5-5742-4067-a953-7d94ac45e415" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.843585 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/339283f2-dc57-4316-a334-6d79fbf2be9b-catalog-content\") pod \"redhat-operators-d6vp9\" (UID: \"339283f2-dc57-4316-a334-6d79fbf2be9b\") " pod="openshift-marketplace/redhat-operators-d6vp9" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.843722 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/339283f2-dc57-4316-a334-6d79fbf2be9b-utilities\") pod \"redhat-operators-d6vp9\" (UID: \"339283f2-dc57-4316-a334-6d79fbf2be9b\") " pod="openshift-marketplace/redhat-operators-d6vp9" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.843760 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw8q2\" (UniqueName: \"kubernetes.io/projected/339283f2-dc57-4316-a334-6d79fbf2be9b-kube-api-access-zw8q2\") pod \"redhat-operators-d6vp9\" (UID: \"339283f2-dc57-4316-a334-6d79fbf2be9b\") " pod="openshift-marketplace/redhat-operators-d6vp9" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.844619 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/339283f2-dc57-4316-a334-6d79fbf2be9b-catalog-content\") pod \"redhat-operators-d6vp9\" (UID: \"339283f2-dc57-4316-a334-6d79fbf2be9b\") " pod="openshift-marketplace/redhat-operators-d6vp9" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.844651 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/339283f2-dc57-4316-a334-6d79fbf2be9b-utilities\") pod \"redhat-operators-d6vp9\" (UID: \"339283f2-dc57-4316-a334-6d79fbf2be9b\") " pod="openshift-marketplace/redhat-operators-d6vp9" Oct 07 13:47:49 crc kubenswrapper[4959]: I1007 13:47:49.872410 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw8q2\" (UniqueName: \"kubernetes.io/projected/339283f2-dc57-4316-a334-6d79fbf2be9b-kube-api-access-zw8q2\") pod \"redhat-operators-d6vp9\" (UID: \"339283f2-dc57-4316-a334-6d79fbf2be9b\") " pod="openshift-marketplace/redhat-operators-d6vp9" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.011501 4959 generic.go:334] "Generic (PLEG): container finished" podID="601096a4-0f0e-4c06-bce3-bdccdd1e370b" containerID="642608cc7d5c158344869ed239420aabbd43174bc508a958b0f238a8a518904b" exitCode=0 Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.011565 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"601096a4-0f0e-4c06-bce3-bdccdd1e370b","Type":"ContainerDied","Data":"642608cc7d5c158344869ed239420aabbd43174bc508a958b0f238a8a518904b"} Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.017764 4959 generic.go:334] "Generic (PLEG): container finished" podID="44b4f9d4-243a-46a8-ab8a-4fb5ae11816b" containerID="08ec0d46e8b05d2c9e1c30ce7036a48dc857a3f3dbe44b2690227f9074441034" exitCode=0 Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.017831 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c94wx" event={"ID":"44b4f9d4-243a-46a8-ab8a-4fb5ae11816b","Type":"ContainerDied","Data":"08ec0d46e8b05d2c9e1c30ce7036a48dc857a3f3dbe44b2690227f9074441034"} Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.018380 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c94wx" event={"ID":"44b4f9d4-243a-46a8-ab8a-4fb5ae11816b","Type":"ContainerStarted","Data":"dee5554320b9e597f40ea1b85d874156840674b24f8986273a37c1e05c47ffac"} Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.022378 4959 generic.go:334] "Generic (PLEG): container finished" podID="11da31c7-46e5-4d9c-b11d-b67fae805f5d" containerID="bdefccd76c25c5312021d53ad897f45a013c7ba5bb4e8ece974659cc95c0b843" exitCode=0 Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.022435 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5brqg" event={"ID":"11da31c7-46e5-4d9c-b11d-b67fae805f5d","Type":"ContainerDied","Data":"bdefccd76c25c5312021d53ad897f45a013c7ba5bb4e8ece974659cc95c0b843"} Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.022468 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5brqg" event={"ID":"11da31c7-46e5-4d9c-b11d-b67fae805f5d","Type":"ContainerStarted","Data":"d36ad2c834f36017d2e03645259f49c7465c493686d70a19d0bf9060e7a03ce1"} Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.025834 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d6vp9" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.051498 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pxlrd"] Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.052661 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pxlrd" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.053469 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330745-wbwkx" event={"ID":"d0a596b0-4162-4b10-8880-3ea0421b8c74","Type":"ContainerDied","Data":"1e512e40764b77383cb5bbfdec10776f2a2dca9062b7365ad69d76968d5aa5d3"} Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.053521 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e512e40764b77383cb5bbfdec10776f2a2dca9062b7365ad69d76968d5aa5d3" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.053845 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330745-wbwkx" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.065485 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pxlrd"] Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.149710 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq8jg\" (UniqueName: \"kubernetes.io/projected/087e1130-9588-4d57-a552-6c868170c308-kube-api-access-sq8jg\") pod \"redhat-operators-pxlrd\" (UID: \"087e1130-9588-4d57-a552-6c868170c308\") " pod="openshift-marketplace/redhat-operators-pxlrd" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.149851 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/087e1130-9588-4d57-a552-6c868170c308-catalog-content\") pod \"redhat-operators-pxlrd\" (UID: \"087e1130-9588-4d57-a552-6c868170c308\") " pod="openshift-marketplace/redhat-operators-pxlrd" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.149947 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/087e1130-9588-4d57-a552-6c868170c308-utilities\") pod \"redhat-operators-pxlrd\" (UID: \"087e1130-9588-4d57-a552-6c868170c308\") " pod="openshift-marketplace/redhat-operators-pxlrd" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.251972 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/087e1130-9588-4d57-a552-6c868170c308-utilities\") pod \"redhat-operators-pxlrd\" (UID: \"087e1130-9588-4d57-a552-6c868170c308\") " pod="openshift-marketplace/redhat-operators-pxlrd" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.252047 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq8jg\" (UniqueName: \"kubernetes.io/projected/087e1130-9588-4d57-a552-6c868170c308-kube-api-access-sq8jg\") pod \"redhat-operators-pxlrd\" (UID: \"087e1130-9588-4d57-a552-6c868170c308\") " pod="openshift-marketplace/redhat-operators-pxlrd" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.252092 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/087e1130-9588-4d57-a552-6c868170c308-catalog-content\") pod \"redhat-operators-pxlrd\" (UID: \"087e1130-9588-4d57-a552-6c868170c308\") " pod="openshift-marketplace/redhat-operators-pxlrd" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.252536 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/087e1130-9588-4d57-a552-6c868170c308-catalog-content\") pod \"redhat-operators-pxlrd\" (UID: \"087e1130-9588-4d57-a552-6c868170c308\") " pod="openshift-marketplace/redhat-operators-pxlrd" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.252604 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/087e1130-9588-4d57-a552-6c868170c308-utilities\") pod \"redhat-operators-pxlrd\" (UID: \"087e1130-9588-4d57-a552-6c868170c308\") " pod="openshift-marketplace/redhat-operators-pxlrd" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.271589 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq8jg\" (UniqueName: \"kubernetes.io/projected/087e1130-9588-4d57-a552-6c868170c308-kube-api-access-sq8jg\") pod \"redhat-operators-pxlrd\" (UID: \"087e1130-9588-4d57-a552-6c868170c308\") " pod="openshift-marketplace/redhat-operators-pxlrd" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.352041 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.352252 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.362142 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.396363 4959 patch_prober.go:28] interesting pod/downloads-7954f5f757-7bxtq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.396384 4959 patch_prober.go:28] interesting pod/downloads-7954f5f757-7bxtq container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.396418 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7bxtq" podUID="8307e459-035b-4c33-bb8e-b5a698618036" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.396429 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-7bxtq" podUID="8307e459-035b-4c33-bb8e-b5a698618036" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.467926 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pxlrd" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.494037 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-xwlzm" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.501895 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.505238 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.508078 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.508386 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.512839 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.526185 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d6vp9"] Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.557678 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/50a8e4e2-c0eb-4fba-ac68-03824eb1f99f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"50a8e4e2-c0eb-4fba-ac68-03824eb1f99f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.557767 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/50a8e4e2-c0eb-4fba-ac68-03824eb1f99f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"50a8e4e2-c0eb-4fba-ac68-03824eb1f99f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.630077 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.630154 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.666953 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/50a8e4e2-c0eb-4fba-ac68-03824eb1f99f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"50a8e4e2-c0eb-4fba-ac68-03824eb1f99f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.689791 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/50a8e4e2-c0eb-4fba-ac68-03824eb1f99f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"50a8e4e2-c0eb-4fba-ac68-03824eb1f99f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.690356 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/50a8e4e2-c0eb-4fba-ac68-03824eb1f99f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"50a8e4e2-c0eb-4fba-ac68-03824eb1f99f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.696530 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/50a8e4e2-c0eb-4fba-ac68-03824eb1f99f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"50a8e4e2-c0eb-4fba-ac68-03824eb1f99f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.776837 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pxlrd"] Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.818700 4959 patch_prober.go:28] interesting pod/router-default-5444994796-8nt7g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 13:47:50 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 07 13:47:50 crc kubenswrapper[4959]: [+]process-running ok Oct 07 13:47:50 crc kubenswrapper[4959]: healthz check failed Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.818797 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nt7g" podUID="67543fb5-5742-4067-a953-7d94ac45e415" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 13:47:50 crc kubenswrapper[4959]: W1007 13:47:50.861211 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod087e1130_9588_4d57_a552_6c868170c308.slice/crio-d71a93a1d58c858fbda701797685b73a253cae5d7ee4d04d61c6c092b368fe87 WatchSource:0}: Error finding container d71a93a1d58c858fbda701797685b73a253cae5d7ee4d04d61c6c092b368fe87: Status 404 returned error can't find the container with id d71a93a1d58c858fbda701797685b73a253cae5d7ee4d04d61c6c092b368fe87 Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.881669 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 13:47:50 crc kubenswrapper[4959]: I1007 13:47:50.916647 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:47:51 crc kubenswrapper[4959]: I1007 13:47:51.091143 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxlrd" event={"ID":"087e1130-9588-4d57-a552-6c868170c308","Type":"ContainerStarted","Data":"d71a93a1d58c858fbda701797685b73a253cae5d7ee4d04d61c6c092b368fe87"} Oct 07 13:47:51 crc kubenswrapper[4959]: I1007 13:47:51.116718 4959 generic.go:334] "Generic (PLEG): container finished" podID="339283f2-dc57-4316-a334-6d79fbf2be9b" containerID="b1ac98632af547caf91e4d584980a4a3bc41aa0808cab60cffc1cb44355c9fbb" exitCode=0 Oct 07 13:47:51 crc kubenswrapper[4959]: I1007 13:47:51.118248 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d6vp9" event={"ID":"339283f2-dc57-4316-a334-6d79fbf2be9b","Type":"ContainerDied","Data":"b1ac98632af547caf91e4d584980a4a3bc41aa0808cab60cffc1cb44355c9fbb"} Oct 07 13:47:51 crc kubenswrapper[4959]: I1007 13:47:51.118284 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d6vp9" event={"ID":"339283f2-dc57-4316-a334-6d79fbf2be9b","Type":"ContainerStarted","Data":"361cf57d8a72a237d4bab5b98cd68a869d83fcef7329840871147fdac46e2121"} Oct 07 13:47:51 crc kubenswrapper[4959]: I1007 13:47:51.125856 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-tv7sf" Oct 07 13:47:51 crc kubenswrapper[4959]: I1007 13:47:51.445710 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 07 13:47:51 crc kubenswrapper[4959]: W1007 13:47:51.499766 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod50a8e4e2_c0eb_4fba_ac68_03824eb1f99f.slice/crio-bc3fb69cd46602a2d0cd32a0b2e7551a4a300324b5a587cd87b34f608111f10f WatchSource:0}: Error finding container bc3fb69cd46602a2d0cd32a0b2e7551a4a300324b5a587cd87b34f608111f10f: Status 404 returned error can't find the container with id bc3fb69cd46602a2d0cd32a0b2e7551a4a300324b5a587cd87b34f608111f10f Oct 07 13:47:51 crc kubenswrapper[4959]: I1007 13:47:51.625809 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 13:47:51 crc kubenswrapper[4959]: I1007 13:47:51.732706 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/601096a4-0f0e-4c06-bce3-bdccdd1e370b-kubelet-dir\") pod \"601096a4-0f0e-4c06-bce3-bdccdd1e370b\" (UID: \"601096a4-0f0e-4c06-bce3-bdccdd1e370b\") " Oct 07 13:47:51 crc kubenswrapper[4959]: I1007 13:47:51.732834 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/601096a4-0f0e-4c06-bce3-bdccdd1e370b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "601096a4-0f0e-4c06-bce3-bdccdd1e370b" (UID: "601096a4-0f0e-4c06-bce3-bdccdd1e370b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 13:47:51 crc kubenswrapper[4959]: I1007 13:47:51.733012 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/601096a4-0f0e-4c06-bce3-bdccdd1e370b-kube-api-access\") pod \"601096a4-0f0e-4c06-bce3-bdccdd1e370b\" (UID: \"601096a4-0f0e-4c06-bce3-bdccdd1e370b\") " Oct 07 13:47:51 crc kubenswrapper[4959]: I1007 13:47:51.733366 4959 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/601096a4-0f0e-4c06-bce3-bdccdd1e370b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 07 13:47:51 crc kubenswrapper[4959]: I1007 13:47:51.739978 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/601096a4-0f0e-4c06-bce3-bdccdd1e370b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "601096a4-0f0e-4c06-bce3-bdccdd1e370b" (UID: "601096a4-0f0e-4c06-bce3-bdccdd1e370b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:47:51 crc kubenswrapper[4959]: I1007 13:47:51.812568 4959 patch_prober.go:28] interesting pod/router-default-5444994796-8nt7g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 13:47:51 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 07 13:47:51 crc kubenswrapper[4959]: [+]process-running ok Oct 07 13:47:51 crc kubenswrapper[4959]: healthz check failed Oct 07 13:47:51 crc kubenswrapper[4959]: I1007 13:47:51.812625 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nt7g" podUID="67543fb5-5742-4067-a953-7d94ac45e415" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 13:47:51 crc kubenswrapper[4959]: I1007 13:47:51.835168 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/601096a4-0f0e-4c06-bce3-bdccdd1e370b-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 07 13:47:52 crc kubenswrapper[4959]: I1007 13:47:52.125279 4959 generic.go:334] "Generic (PLEG): container finished" podID="087e1130-9588-4d57-a552-6c868170c308" containerID="04e934f377283b5ed1b5e8feacc637ff2e4499fa1317874e293f19b9d4dc3651" exitCode=0 Oct 07 13:47:52 crc kubenswrapper[4959]: I1007 13:47:52.125331 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxlrd" event={"ID":"087e1130-9588-4d57-a552-6c868170c308","Type":"ContainerDied","Data":"04e934f377283b5ed1b5e8feacc637ff2e4499fa1317874e293f19b9d4dc3651"} Oct 07 13:47:52 crc kubenswrapper[4959]: I1007 13:47:52.139593 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"50a8e4e2-c0eb-4fba-ac68-03824eb1f99f","Type":"ContainerStarted","Data":"bc3fb69cd46602a2d0cd32a0b2e7551a4a300324b5a587cd87b34f608111f10f"} Oct 07 13:47:52 crc kubenswrapper[4959]: I1007 13:47:52.146595 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 07 13:47:52 crc kubenswrapper[4959]: I1007 13:47:52.146649 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"601096a4-0f0e-4c06-bce3-bdccdd1e370b","Type":"ContainerDied","Data":"56657ce5bc739a96605641045ceb210b12af55764bac692c91f2e835611cfef7"} Oct 07 13:47:52 crc kubenswrapper[4959]: I1007 13:47:52.146674 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56657ce5bc739a96605641045ceb210b12af55764bac692c91f2e835611cfef7" Oct 07 13:47:52 crc kubenswrapper[4959]: I1007 13:47:52.811010 4959 patch_prober.go:28] interesting pod/router-default-5444994796-8nt7g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 13:47:52 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 07 13:47:52 crc kubenswrapper[4959]: [+]process-running ok Oct 07 13:47:52 crc kubenswrapper[4959]: healthz check failed Oct 07 13:47:52 crc kubenswrapper[4959]: I1007 13:47:52.811284 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nt7g" podUID="67543fb5-5742-4067-a953-7d94ac45e415" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 13:47:53 crc kubenswrapper[4959]: I1007 13:47:53.191733 4959 generic.go:334] "Generic (PLEG): container finished" podID="50a8e4e2-c0eb-4fba-ac68-03824eb1f99f" containerID="d5d92d8976b07d7a80d5b46bc20950797a62eb674326f8dd4d53ff5ce904e6c8" exitCode=0 Oct 07 13:47:53 crc kubenswrapper[4959]: I1007 13:47:53.191828 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"50a8e4e2-c0eb-4fba-ac68-03824eb1f99f","Type":"ContainerDied","Data":"d5d92d8976b07d7a80d5b46bc20950797a62eb674326f8dd4d53ff5ce904e6c8"} Oct 07 13:47:53 crc kubenswrapper[4959]: I1007 13:47:53.810810 4959 patch_prober.go:28] interesting pod/router-default-5444994796-8nt7g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 13:47:53 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 07 13:47:53 crc kubenswrapper[4959]: [+]process-running ok Oct 07 13:47:53 crc kubenswrapper[4959]: healthz check failed Oct 07 13:47:53 crc kubenswrapper[4959]: I1007 13:47:53.810951 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nt7g" podUID="67543fb5-5742-4067-a953-7d94ac45e415" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 13:47:54 crc kubenswrapper[4959]: I1007 13:47:54.811555 4959 patch_prober.go:28] interesting pod/router-default-5444994796-8nt7g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 13:47:54 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 07 13:47:54 crc kubenswrapper[4959]: [+]process-running ok Oct 07 13:47:54 crc kubenswrapper[4959]: healthz check failed Oct 07 13:47:54 crc kubenswrapper[4959]: I1007 13:47:54.812193 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nt7g" podUID="67543fb5-5742-4067-a953-7d94ac45e415" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 13:47:55 crc kubenswrapper[4959]: I1007 13:47:55.576401 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-rfcqs" Oct 07 13:47:55 crc kubenswrapper[4959]: I1007 13:47:55.816841 4959 patch_prober.go:28] interesting pod/router-default-5444994796-8nt7g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 13:47:55 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 07 13:47:55 crc kubenswrapper[4959]: [+]process-running ok Oct 07 13:47:55 crc kubenswrapper[4959]: healthz check failed Oct 07 13:47:55 crc kubenswrapper[4959]: I1007 13:47:55.816910 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nt7g" podUID="67543fb5-5742-4067-a953-7d94ac45e415" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 13:47:56 crc kubenswrapper[4959]: I1007 13:47:56.809390 4959 patch_prober.go:28] interesting pod/router-default-5444994796-8nt7g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 13:47:56 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 07 13:47:56 crc kubenswrapper[4959]: [+]process-running ok Oct 07 13:47:56 crc kubenswrapper[4959]: healthz check failed Oct 07 13:47:56 crc kubenswrapper[4959]: I1007 13:47:56.809478 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nt7g" podUID="67543fb5-5742-4067-a953-7d94ac45e415" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 13:47:57 crc kubenswrapper[4959]: I1007 13:47:57.809898 4959 patch_prober.go:28] interesting pod/router-default-5444994796-8nt7g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 07 13:47:57 crc kubenswrapper[4959]: [+]has-synced ok Oct 07 13:47:57 crc kubenswrapper[4959]: [+]process-running ok Oct 07 13:47:57 crc kubenswrapper[4959]: healthz check failed Oct 07 13:47:57 crc kubenswrapper[4959]: I1007 13:47:57.810376 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nt7g" podUID="67543fb5-5742-4067-a953-7d94ac45e415" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 07 13:47:58 crc kubenswrapper[4959]: I1007 13:47:58.809196 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-8nt7g" Oct 07 13:47:58 crc kubenswrapper[4959]: I1007 13:47:58.811261 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-8nt7g" Oct 07 13:47:59 crc kubenswrapper[4959]: I1007 13:47:59.801493 4959 patch_prober.go:28] interesting pod/console-f9d7485db-pcvlp container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.19:8443/health\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Oct 07 13:47:59 crc kubenswrapper[4959]: I1007 13:47:59.803502 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-pcvlp" podUID="4fca2a0e-545f-43b0-b4ec-e5f15babcd71" containerName="console" probeResult="failure" output="Get \"https://10.217.0.19:8443/health\": dial tcp 10.217.0.19:8443: connect: connection refused" Oct 07 13:48:00 crc kubenswrapper[4959]: I1007 13:48:00.398069 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-7bxtq" Oct 07 13:48:01 crc kubenswrapper[4959]: I1007 13:48:01.448024 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 13:48:01 crc kubenswrapper[4959]: I1007 13:48:01.603473 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/50a8e4e2-c0eb-4fba-ac68-03824eb1f99f-kubelet-dir\") pod \"50a8e4e2-c0eb-4fba-ac68-03824eb1f99f\" (UID: \"50a8e4e2-c0eb-4fba-ac68-03824eb1f99f\") " Oct 07 13:48:01 crc kubenswrapper[4959]: I1007 13:48:01.603571 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/50a8e4e2-c0eb-4fba-ac68-03824eb1f99f-kube-api-access\") pod \"50a8e4e2-c0eb-4fba-ac68-03824eb1f99f\" (UID: \"50a8e4e2-c0eb-4fba-ac68-03824eb1f99f\") " Oct 07 13:48:01 crc kubenswrapper[4959]: I1007 13:48:01.603589 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/50a8e4e2-c0eb-4fba-ac68-03824eb1f99f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "50a8e4e2-c0eb-4fba-ac68-03824eb1f99f" (UID: "50a8e4e2-c0eb-4fba-ac68-03824eb1f99f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 13:48:01 crc kubenswrapper[4959]: I1007 13:48:01.603840 4959 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/50a8e4e2-c0eb-4fba-ac68-03824eb1f99f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 07 13:48:01 crc kubenswrapper[4959]: I1007 13:48:01.616406 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50a8e4e2-c0eb-4fba-ac68-03824eb1f99f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "50a8e4e2-c0eb-4fba-ac68-03824eb1f99f" (UID: "50a8e4e2-c0eb-4fba-ac68-03824eb1f99f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:48:01 crc kubenswrapper[4959]: I1007 13:48:01.705352 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/50a8e4e2-c0eb-4fba-ac68-03824eb1f99f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 07 13:48:02 crc kubenswrapper[4959]: I1007 13:48:02.282079 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"50a8e4e2-c0eb-4fba-ac68-03824eb1f99f","Type":"ContainerDied","Data":"bc3fb69cd46602a2d0cd32a0b2e7551a4a300324b5a587cd87b34f608111f10f"} Oct 07 13:48:02 crc kubenswrapper[4959]: I1007 13:48:02.282142 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc3fb69cd46602a2d0cd32a0b2e7551a4a300324b5a587cd87b34f608111f10f" Oct 07 13:48:02 crc kubenswrapper[4959]: I1007 13:48:02.282199 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 07 13:48:06 crc kubenswrapper[4959]: I1007 13:48:06.574739 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-metrics-certs\") pod \"network-metrics-daemon-zdps5\" (UID: \"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\") " pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:48:06 crc kubenswrapper[4959]: I1007 13:48:06.588705 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50d4ac5c-0a16-4530-8500-9975eb2b6b5a-metrics-certs\") pod \"network-metrics-daemon-zdps5\" (UID: \"50d4ac5c-0a16-4530-8500-9975eb2b6b5a\") " pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:48:06 crc kubenswrapper[4959]: I1007 13:48:06.593609 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zdps5" Oct 07 13:48:07 crc kubenswrapper[4959]: I1007 13:48:07.319736 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:48:09 crc kubenswrapper[4959]: I1007 13:48:09.807981 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-pcvlp" Oct 07 13:48:09 crc kubenswrapper[4959]: I1007 13:48:09.814644 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-pcvlp" Oct 07 13:48:20 crc kubenswrapper[4959]: I1007 13:48:20.164581 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-94zbw" Oct 07 13:48:20 crc kubenswrapper[4959]: I1007 13:48:20.630414 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 13:48:20 crc kubenswrapper[4959]: I1007 13:48:20.630495 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 13:48:22 crc kubenswrapper[4959]: E1007 13:48:22.119778 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 07 13:48:22 crc kubenswrapper[4959]: E1007 13:48:22.120045 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zw8q2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-d6vp9_openshift-marketplace(339283f2-dc57-4316-a334-6d79fbf2be9b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 07 13:48:22 crc kubenswrapper[4959]: E1007 13:48:22.121272 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-d6vp9" podUID="339283f2-dc57-4316-a334-6d79fbf2be9b" Oct 07 13:48:23 crc kubenswrapper[4959]: E1007 13:48:23.365685 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-d6vp9" podUID="339283f2-dc57-4316-a334-6d79fbf2be9b" Oct 07 13:48:23 crc kubenswrapper[4959]: E1007 13:48:23.453476 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 07 13:48:23 crc kubenswrapper[4959]: E1007 13:48:23.453716 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rw9ct,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-9bkvf_openshift-marketplace(f0e27f2e-7704-4082-91e5-571830f218fc): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 07 13:48:23 crc kubenswrapper[4959]: E1007 13:48:23.455069 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-9bkvf" podUID="f0e27f2e-7704-4082-91e5-571830f218fc" Oct 07 13:48:23 crc kubenswrapper[4959]: E1007 13:48:23.467578 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 07 13:48:23 crc kubenswrapper[4959]: E1007 13:48:23.467724 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qhdvl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-5r424_openshift-marketplace(4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 07 13:48:23 crc kubenswrapper[4959]: E1007 13:48:23.469158 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-5r424" podUID="4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac" Oct 07 13:48:24 crc kubenswrapper[4959]: E1007 13:48:24.515012 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-5r424" podUID="4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac" Oct 07 13:48:24 crc kubenswrapper[4959]: E1007 13:48:24.515465 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-9bkvf" podUID="f0e27f2e-7704-4082-91e5-571830f218fc" Oct 07 13:48:24 crc kubenswrapper[4959]: E1007 13:48:24.612496 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 07 13:48:24 crc kubenswrapper[4959]: E1007 13:48:24.612686 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-grzcm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-kkxdr_openshift-marketplace(a5295901-596b-447c-8d7a-bac77b140153): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 07 13:48:24 crc kubenswrapper[4959]: E1007 13:48:24.613834 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-kkxdr" podUID="a5295901-596b-447c-8d7a-bac77b140153" Oct 07 13:48:24 crc kubenswrapper[4959]: E1007 13:48:24.657791 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 07 13:48:24 crc kubenswrapper[4959]: E1007 13:48:24.657940 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sq8jg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-pxlrd_openshift-marketplace(087e1130-9588-4d57-a552-6c868170c308): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 07 13:48:24 crc kubenswrapper[4959]: E1007 13:48:24.659060 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-pxlrd" podUID="087e1130-9588-4d57-a552-6c868170c308" Oct 07 13:48:26 crc kubenswrapper[4959]: E1007 13:48:26.565238 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-pxlrd" podUID="087e1130-9588-4d57-a552-6c868170c308" Oct 07 13:48:26 crc kubenswrapper[4959]: E1007 13:48:26.566659 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-kkxdr" podUID="a5295901-596b-447c-8d7a-bac77b140153" Oct 07 13:48:26 crc kubenswrapper[4959]: E1007 13:48:26.633401 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 07 13:48:26 crc kubenswrapper[4959]: E1007 13:48:26.633604 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7vvdz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-t7mrz_openshift-marketplace(ebb5441f-839a-408b-bd51-3460006c3d83): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 07 13:48:26 crc kubenswrapper[4959]: E1007 13:48:26.635136 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-t7mrz" podUID="ebb5441f-839a-408b-bd51-3460006c3d83" Oct 07 13:48:27 crc kubenswrapper[4959]: I1007 13:48:27.085973 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 07 13:48:27 crc kubenswrapper[4959]: E1007 13:48:27.227385 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 07 13:48:27 crc kubenswrapper[4959]: E1007 13:48:27.227553 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tk75g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-5brqg_openshift-marketplace(11da31c7-46e5-4d9c-b11d-b67fae805f5d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 07 13:48:27 crc kubenswrapper[4959]: E1007 13:48:27.229242 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-5brqg" podUID="11da31c7-46e5-4d9c-b11d-b67fae805f5d" Oct 07 13:48:27 crc kubenswrapper[4959]: E1007 13:48:27.280688 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 07 13:48:27 crc kubenswrapper[4959]: E1007 13:48:27.281240 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-925d4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-c94wx_openshift-marketplace(44b4f9d4-243a-46a8-ab8a-4fb5ae11816b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 07 13:48:27 crc kubenswrapper[4959]: E1007 13:48:27.282948 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-c94wx" podUID="44b4f9d4-243a-46a8-ab8a-4fb5ae11816b" Oct 07 13:48:27 crc kubenswrapper[4959]: E1007 13:48:27.438630 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-5brqg" podUID="11da31c7-46e5-4d9c-b11d-b67fae805f5d" Oct 07 13:48:27 crc kubenswrapper[4959]: E1007 13:48:27.438660 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-c94wx" podUID="44b4f9d4-243a-46a8-ab8a-4fb5ae11816b" Oct 07 13:48:27 crc kubenswrapper[4959]: E1007 13:48:27.439012 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-t7mrz" podUID="ebb5441f-839a-408b-bd51-3460006c3d83" Oct 07 13:48:27 crc kubenswrapper[4959]: I1007 13:48:27.582794 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-zdps5"] Oct 07 13:48:27 crc kubenswrapper[4959]: W1007 13:48:27.587437 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50d4ac5c_0a16_4530_8500_9975eb2b6b5a.slice/crio-04f221c36c5a3a7ce12f0564727fd7210ee5e247b987ebd1d934531297ead241 WatchSource:0}: Error finding container 04f221c36c5a3a7ce12f0564727fd7210ee5e247b987ebd1d934531297ead241: Status 404 returned error can't find the container with id 04f221c36c5a3a7ce12f0564727fd7210ee5e247b987ebd1d934531297ead241 Oct 07 13:48:28 crc kubenswrapper[4959]: I1007 13:48:28.446909 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zdps5" event={"ID":"50d4ac5c-0a16-4530-8500-9975eb2b6b5a","Type":"ContainerStarted","Data":"1cefbb9477811a00f20329c9b81a02bbc0de07c9b356d4a6a599915bb7641598"} Oct 07 13:48:28 crc kubenswrapper[4959]: I1007 13:48:28.447327 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zdps5" event={"ID":"50d4ac5c-0a16-4530-8500-9975eb2b6b5a","Type":"ContainerStarted","Data":"3ab1ec86daa0b4d1effc37e3508a8e38d68c43d82bdd4898d2b7b9b6315e272a"} Oct 07 13:48:28 crc kubenswrapper[4959]: I1007 13:48:28.447345 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zdps5" event={"ID":"50d4ac5c-0a16-4530-8500-9975eb2b6b5a","Type":"ContainerStarted","Data":"04f221c36c5a3a7ce12f0564727fd7210ee5e247b987ebd1d934531297ead241"} Oct 07 13:48:28 crc kubenswrapper[4959]: I1007 13:48:28.466685 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-zdps5" podStartSLOduration=164.466656521 podStartE2EDuration="2m44.466656521s" podCreationTimestamp="2025-10-07 13:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:48:28.461081027 +0000 UTC m=+190.544485372" watchObservedRunningTime="2025-10-07 13:48:28.466656521 +0000 UTC m=+190.550060856" Oct 07 13:48:49 crc kubenswrapper[4959]: I1007 13:48:49.607145 4959 generic.go:334] "Generic (PLEG): container finished" podID="f0e27f2e-7704-4082-91e5-571830f218fc" containerID="a529d7542c01271a08ed640b28c836c3f83759c5ed3bf7f2e275b697003efa1d" exitCode=0 Oct 07 13:48:49 crc kubenswrapper[4959]: I1007 13:48:49.607199 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9bkvf" event={"ID":"f0e27f2e-7704-4082-91e5-571830f218fc","Type":"ContainerDied","Data":"a529d7542c01271a08ed640b28c836c3f83759c5ed3bf7f2e275b697003efa1d"} Oct 07 13:48:49 crc kubenswrapper[4959]: I1007 13:48:49.612421 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5brqg" event={"ID":"11da31c7-46e5-4d9c-b11d-b67fae805f5d","Type":"ContainerStarted","Data":"1321ae2985977e169b34124567d766faa3d33170d43f74a6d9f0cec83dd81130"} Oct 07 13:48:49 crc kubenswrapper[4959]: I1007 13:48:49.618685 4959 generic.go:334] "Generic (PLEG): container finished" podID="339283f2-dc57-4316-a334-6d79fbf2be9b" containerID="00a246533cb95ee3510722163a228aba26a544d17b441a6cb7e496d9ea99c053" exitCode=0 Oct 07 13:48:49 crc kubenswrapper[4959]: I1007 13:48:49.618740 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d6vp9" event={"ID":"339283f2-dc57-4316-a334-6d79fbf2be9b","Type":"ContainerDied","Data":"00a246533cb95ee3510722163a228aba26a544d17b441a6cb7e496d9ea99c053"} Oct 07 13:48:49 crc kubenswrapper[4959]: I1007 13:48:49.627953 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkxdr" event={"ID":"a5295901-596b-447c-8d7a-bac77b140153","Type":"ContainerStarted","Data":"1fe1aa977e33c0e9309d7891c1abf7026211722775c574c75004397def0274f3"} Oct 07 13:48:49 crc kubenswrapper[4959]: I1007 13:48:49.637859 4959 generic.go:334] "Generic (PLEG): container finished" podID="4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac" containerID="0d0836308b8ddf7f378e0a545fe3d6f67656e86ed37c3ba57f5fa83974e94e92" exitCode=0 Oct 07 13:48:49 crc kubenswrapper[4959]: I1007 13:48:49.637896 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5r424" event={"ID":"4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac","Type":"ContainerDied","Data":"0d0836308b8ddf7f378e0a545fe3d6f67656e86ed37c3ba57f5fa83974e94e92"} Oct 07 13:48:50 crc kubenswrapper[4959]: I1007 13:48:50.630646 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 13:48:50 crc kubenswrapper[4959]: I1007 13:48:50.631415 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 13:48:50 crc kubenswrapper[4959]: I1007 13:48:50.631480 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 13:48:50 crc kubenswrapper[4959]: I1007 13:48:50.632325 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f"} pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 13:48:50 crc kubenswrapper[4959]: I1007 13:48:50.632637 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" containerID="cri-o://bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f" gracePeriod=600 Oct 07 13:48:50 crc kubenswrapper[4959]: I1007 13:48:50.648963 4959 generic.go:334] "Generic (PLEG): container finished" podID="44b4f9d4-243a-46a8-ab8a-4fb5ae11816b" containerID="cd25de8a142f807182b6ed2b264dfffe32efe9fba00f41edc369445587a549da" exitCode=0 Oct 07 13:48:50 crc kubenswrapper[4959]: I1007 13:48:50.649034 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c94wx" event={"ID":"44b4f9d4-243a-46a8-ab8a-4fb5ae11816b","Type":"ContainerDied","Data":"cd25de8a142f807182b6ed2b264dfffe32efe9fba00f41edc369445587a549da"} Oct 07 13:48:50 crc kubenswrapper[4959]: I1007 13:48:50.662745 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5r424" event={"ID":"4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac","Type":"ContainerStarted","Data":"44ca7c831bca81e37b93e090d7dd71055c8d6605796726e271b35d43b47d707a"} Oct 07 13:48:50 crc kubenswrapper[4959]: I1007 13:48:50.666868 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9bkvf" event={"ID":"f0e27f2e-7704-4082-91e5-571830f218fc","Type":"ContainerStarted","Data":"7e53afb61245f016c470ea7180203eefef643a82a67cf7df3e2bfe4b44f7becd"} Oct 07 13:48:50 crc kubenswrapper[4959]: I1007 13:48:50.669569 4959 generic.go:334] "Generic (PLEG): container finished" podID="087e1130-9588-4d57-a552-6c868170c308" containerID="f87747fb36379f2e7835619556bc4e51bc21be04b9349d85ffdb200b8020ffd2" exitCode=0 Oct 07 13:48:50 crc kubenswrapper[4959]: I1007 13:48:50.669639 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxlrd" event={"ID":"087e1130-9588-4d57-a552-6c868170c308","Type":"ContainerDied","Data":"f87747fb36379f2e7835619556bc4e51bc21be04b9349d85ffdb200b8020ffd2"} Oct 07 13:48:50 crc kubenswrapper[4959]: I1007 13:48:50.671816 4959 generic.go:334] "Generic (PLEG): container finished" podID="11da31c7-46e5-4d9c-b11d-b67fae805f5d" containerID="1321ae2985977e169b34124567d766faa3d33170d43f74a6d9f0cec83dd81130" exitCode=0 Oct 07 13:48:50 crc kubenswrapper[4959]: I1007 13:48:50.671862 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5brqg" event={"ID":"11da31c7-46e5-4d9c-b11d-b67fae805f5d","Type":"ContainerDied","Data":"1321ae2985977e169b34124567d766faa3d33170d43f74a6d9f0cec83dd81130"} Oct 07 13:48:50 crc kubenswrapper[4959]: I1007 13:48:50.673788 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d6vp9" event={"ID":"339283f2-dc57-4316-a334-6d79fbf2be9b","Type":"ContainerStarted","Data":"2daa6f510281017ba2c87047c2b8f0b82762823189e3d76885a628912d0e67a0"} Oct 07 13:48:50 crc kubenswrapper[4959]: I1007 13:48:50.679254 4959 generic.go:334] "Generic (PLEG): container finished" podID="a5295901-596b-447c-8d7a-bac77b140153" containerID="1fe1aa977e33c0e9309d7891c1abf7026211722775c574c75004397def0274f3" exitCode=0 Oct 07 13:48:50 crc kubenswrapper[4959]: I1007 13:48:50.679309 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkxdr" event={"ID":"a5295901-596b-447c-8d7a-bac77b140153","Type":"ContainerDied","Data":"1fe1aa977e33c0e9309d7891c1abf7026211722775c574c75004397def0274f3"} Oct 07 13:48:50 crc kubenswrapper[4959]: I1007 13:48:50.682328 4959 generic.go:334] "Generic (PLEG): container finished" podID="ebb5441f-839a-408b-bd51-3460006c3d83" containerID="109edd1598c60b297dfb00054814dabc127bddcf64ba1f23423a35004fa1fc76" exitCode=0 Oct 07 13:48:50 crc kubenswrapper[4959]: I1007 13:48:50.682351 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7mrz" event={"ID":"ebb5441f-839a-408b-bd51-3460006c3d83","Type":"ContainerDied","Data":"109edd1598c60b297dfb00054814dabc127bddcf64ba1f23423a35004fa1fc76"} Oct 07 13:48:50 crc kubenswrapper[4959]: I1007 13:48:50.698338 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5r424" podStartSLOduration=2.543748956 podStartE2EDuration="1m4.698315056s" podCreationTimestamp="2025-10-07 13:47:46 +0000 UTC" firstStartedPulling="2025-10-07 13:47:47.935344054 +0000 UTC m=+150.018748379" lastFinishedPulling="2025-10-07 13:48:50.089910154 +0000 UTC m=+212.173314479" observedRunningTime="2025-10-07 13:48:50.691411284 +0000 UTC m=+212.774815619" watchObservedRunningTime="2025-10-07 13:48:50.698315056 +0000 UTC m=+212.781719401" Oct 07 13:48:50 crc kubenswrapper[4959]: I1007 13:48:50.712903 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9bkvf" podStartSLOduration=3.530500067 podStartE2EDuration="1m4.712879763s" podCreationTimestamp="2025-10-07 13:47:46 +0000 UTC" firstStartedPulling="2025-10-07 13:47:49.003373591 +0000 UTC m=+151.086777916" lastFinishedPulling="2025-10-07 13:48:50.185753287 +0000 UTC m=+212.269157612" observedRunningTime="2025-10-07 13:48:50.710414134 +0000 UTC m=+212.793818469" watchObservedRunningTime="2025-10-07 13:48:50.712879763 +0000 UTC m=+212.796284108" Oct 07 13:48:50 crc kubenswrapper[4959]: I1007 13:48:50.811647 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-d6vp9" podStartSLOduration=2.784204827 podStartE2EDuration="1m1.811623537s" podCreationTimestamp="2025-10-07 13:47:49 +0000 UTC" firstStartedPulling="2025-10-07 13:47:51.124265857 +0000 UTC m=+153.207670182" lastFinishedPulling="2025-10-07 13:48:50.151684567 +0000 UTC m=+212.235088892" observedRunningTime="2025-10-07 13:48:50.811543555 +0000 UTC m=+212.894947900" watchObservedRunningTime="2025-10-07 13:48:50.811623537 +0000 UTC m=+212.895027862" Oct 07 13:48:51 crc kubenswrapper[4959]: I1007 13:48:51.688396 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxlrd" event={"ID":"087e1130-9588-4d57-a552-6c868170c308","Type":"ContainerStarted","Data":"009ef1660013fd06e51f92374b03c12ffbdf04b5f91932b988e61bdb6b125d59"} Oct 07 13:48:51 crc kubenswrapper[4959]: I1007 13:48:51.690119 4959 generic.go:334] "Generic (PLEG): container finished" podID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerID="bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f" exitCode=0 Oct 07 13:48:51 crc kubenswrapper[4959]: I1007 13:48:51.690167 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerDied","Data":"bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f"} Oct 07 13:48:51 crc kubenswrapper[4959]: I1007 13:48:51.690187 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerStarted","Data":"ff34b703f977911ce787f57286ec64739600838a5ec5892759fc7837be39f505"} Oct 07 13:48:51 crc kubenswrapper[4959]: I1007 13:48:51.692029 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5brqg" event={"ID":"11da31c7-46e5-4d9c-b11d-b67fae805f5d","Type":"ContainerStarted","Data":"ba8ee5e3680f017d6ab7b3d17688f06fb81b9f304f066d1cf0cde72c5539f6d9"} Oct 07 13:48:51 crc kubenswrapper[4959]: I1007 13:48:51.694041 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkxdr" event={"ID":"a5295901-596b-447c-8d7a-bac77b140153","Type":"ContainerStarted","Data":"8516b43df0c7fb6fe47ee62563ae339665cc2a77ba523d495be9955bf725b64b"} Oct 07 13:48:51 crc kubenswrapper[4959]: I1007 13:48:51.696761 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7mrz" event={"ID":"ebb5441f-839a-408b-bd51-3460006c3d83","Type":"ContainerStarted","Data":"b2384da6a2e7f871634616e179a6b68fbbec8011b37cc0931b7babc6c27c9546"} Oct 07 13:48:51 crc kubenswrapper[4959]: I1007 13:48:51.699025 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c94wx" event={"ID":"44b4f9d4-243a-46a8-ab8a-4fb5ae11816b","Type":"ContainerStarted","Data":"2948cb9150d3ce5ef0e539f41a521dbe6e02828b34f4ea21d9bb12593c8e4f6f"} Oct 07 13:48:51 crc kubenswrapper[4959]: I1007 13:48:51.710359 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pxlrd" podStartSLOduration=2.648215986 podStartE2EDuration="1m1.710337258s" podCreationTimestamp="2025-10-07 13:47:50 +0000 UTC" firstStartedPulling="2025-10-07 13:47:52.141393178 +0000 UTC m=+154.224797503" lastFinishedPulling="2025-10-07 13:48:51.20351445 +0000 UTC m=+213.286918775" observedRunningTime="2025-10-07 13:48:51.709286159 +0000 UTC m=+213.792690484" watchObservedRunningTime="2025-10-07 13:48:51.710337258 +0000 UTC m=+213.793741583" Oct 07 13:48:51 crc kubenswrapper[4959]: I1007 13:48:51.735795 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5brqg" podStartSLOduration=2.527840284 podStartE2EDuration="1m3.735776448s" podCreationTimestamp="2025-10-07 13:47:48 +0000 UTC" firstStartedPulling="2025-10-07 13:47:50.026446447 +0000 UTC m=+152.109850772" lastFinishedPulling="2025-10-07 13:48:51.234382611 +0000 UTC m=+213.317786936" observedRunningTime="2025-10-07 13:48:51.734541274 +0000 UTC m=+213.817945599" watchObservedRunningTime="2025-10-07 13:48:51.735776448 +0000 UTC m=+213.819180773" Oct 07 13:48:51 crc kubenswrapper[4959]: I1007 13:48:51.763925 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c94wx" podStartSLOduration=2.634372348 podStartE2EDuration="1m3.763906123s" podCreationTimestamp="2025-10-07 13:47:48 +0000 UTC" firstStartedPulling="2025-10-07 13:47:50.020330468 +0000 UTC m=+152.103734793" lastFinishedPulling="2025-10-07 13:48:51.149864243 +0000 UTC m=+213.233268568" observedRunningTime="2025-10-07 13:48:51.763601314 +0000 UTC m=+213.847005659" watchObservedRunningTime="2025-10-07 13:48:51.763906123 +0000 UTC m=+213.847310448" Oct 07 13:48:51 crc kubenswrapper[4959]: I1007 13:48:51.810865 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kkxdr" podStartSLOduration=2.64813618 podStartE2EDuration="1m4.810846522s" podCreationTimestamp="2025-10-07 13:47:47 +0000 UTC" firstStartedPulling="2025-10-07 13:47:48.959990063 +0000 UTC m=+151.043394388" lastFinishedPulling="2025-10-07 13:48:51.122700405 +0000 UTC m=+213.206104730" observedRunningTime="2025-10-07 13:48:51.810244425 +0000 UTC m=+213.893648760" watchObservedRunningTime="2025-10-07 13:48:51.810846522 +0000 UTC m=+213.894250847" Oct 07 13:48:51 crc kubenswrapper[4959]: I1007 13:48:51.832405 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t7mrz" podStartSLOduration=3.527010437 podStartE2EDuration="1m5.832387893s" podCreationTimestamp="2025-10-07 13:47:46 +0000 UTC" firstStartedPulling="2025-10-07 13:47:48.973792834 +0000 UTC m=+151.057197159" lastFinishedPulling="2025-10-07 13:48:51.27917029 +0000 UTC m=+213.362574615" observedRunningTime="2025-10-07 13:48:51.82904529 +0000 UTC m=+213.912449615" watchObservedRunningTime="2025-10-07 13:48:51.832387893 +0000 UTC m=+213.915792208" Oct 07 13:48:54 crc kubenswrapper[4959]: I1007 13:48:54.718729 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-67jl7"] Oct 07 13:48:56 crc kubenswrapper[4959]: I1007 13:48:56.802362 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5r424" Oct 07 13:48:56 crc kubenswrapper[4959]: I1007 13:48:56.802664 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5r424" Oct 07 13:48:56 crc kubenswrapper[4959]: I1007 13:48:56.958622 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5r424" Oct 07 13:48:56 crc kubenswrapper[4959]: I1007 13:48:56.984554 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t7mrz" Oct 07 13:48:56 crc kubenswrapper[4959]: I1007 13:48:56.984600 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-t7mrz" Oct 07 13:48:57 crc kubenswrapper[4959]: I1007 13:48:57.026480 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t7mrz" Oct 07 13:48:57 crc kubenswrapper[4959]: I1007 13:48:57.182651 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9bkvf" Oct 07 13:48:57 crc kubenswrapper[4959]: I1007 13:48:57.182743 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9bkvf" Oct 07 13:48:57 crc kubenswrapper[4959]: I1007 13:48:57.215753 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9bkvf" Oct 07 13:48:57 crc kubenswrapper[4959]: I1007 13:48:57.391171 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kkxdr" Oct 07 13:48:57 crc kubenswrapper[4959]: I1007 13:48:57.391244 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kkxdr" Oct 07 13:48:57 crc kubenswrapper[4959]: I1007 13:48:57.426064 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kkxdr" Oct 07 13:48:57 crc kubenswrapper[4959]: I1007 13:48:57.766411 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kkxdr" Oct 07 13:48:57 crc kubenswrapper[4959]: I1007 13:48:57.777246 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5r424" Oct 07 13:48:57 crc kubenswrapper[4959]: I1007 13:48:57.783090 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9bkvf" Oct 07 13:48:57 crc kubenswrapper[4959]: I1007 13:48:57.786524 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t7mrz" Oct 07 13:48:58 crc kubenswrapper[4959]: I1007 13:48:58.772470 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c94wx" Oct 07 13:48:58 crc kubenswrapper[4959]: I1007 13:48:58.772534 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c94wx" Oct 07 13:48:58 crc kubenswrapper[4959]: I1007 13:48:58.808976 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c94wx" Oct 07 13:48:59 crc kubenswrapper[4959]: I1007 13:48:59.071065 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9bkvf"] Oct 07 13:48:59 crc kubenswrapper[4959]: I1007 13:48:59.166056 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5brqg" Oct 07 13:48:59 crc kubenswrapper[4959]: I1007 13:48:59.166129 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5brqg" Oct 07 13:48:59 crc kubenswrapper[4959]: I1007 13:48:59.206356 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5brqg" Oct 07 13:48:59 crc kubenswrapper[4959]: I1007 13:48:59.269080 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kkxdr"] Oct 07 13:48:59 crc kubenswrapper[4959]: I1007 13:48:59.737670 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kkxdr" podUID="a5295901-596b-447c-8d7a-bac77b140153" containerName="registry-server" containerID="cri-o://8516b43df0c7fb6fe47ee62563ae339665cc2a77ba523d495be9955bf725b64b" gracePeriod=2 Oct 07 13:48:59 crc kubenswrapper[4959]: I1007 13:48:59.782715 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c94wx" Oct 07 13:48:59 crc kubenswrapper[4959]: I1007 13:48:59.784425 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5brqg" Oct 07 13:49:00 crc kubenswrapper[4959]: I1007 13:49:00.026860 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-d6vp9" Oct 07 13:49:00 crc kubenswrapper[4959]: I1007 13:49:00.026900 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-d6vp9" Oct 07 13:49:00 crc kubenswrapper[4959]: I1007 13:49:00.061045 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-d6vp9" Oct 07 13:49:00 crc kubenswrapper[4959]: I1007 13:49:00.468562 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pxlrd" Oct 07 13:49:00 crc kubenswrapper[4959]: I1007 13:49:00.468608 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pxlrd" Oct 07 13:49:00 crc kubenswrapper[4959]: I1007 13:49:00.536616 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pxlrd" Oct 07 13:49:00 crc kubenswrapper[4959]: I1007 13:49:00.741456 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9bkvf" podUID="f0e27f2e-7704-4082-91e5-571830f218fc" containerName="registry-server" containerID="cri-o://7e53afb61245f016c470ea7180203eefef643a82a67cf7df3e2bfe4b44f7becd" gracePeriod=2 Oct 07 13:49:00 crc kubenswrapper[4959]: I1007 13:49:00.774698 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-d6vp9" Oct 07 13:49:00 crc kubenswrapper[4959]: I1007 13:49:00.778472 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pxlrd" Oct 07 13:49:01 crc kubenswrapper[4959]: I1007 13:49:01.471061 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5brqg"] Oct 07 13:49:01 crc kubenswrapper[4959]: I1007 13:49:01.746718 4959 generic.go:334] "Generic (PLEG): container finished" podID="a5295901-596b-447c-8d7a-bac77b140153" containerID="8516b43df0c7fb6fe47ee62563ae339665cc2a77ba523d495be9955bf725b64b" exitCode=0 Oct 07 13:49:01 crc kubenswrapper[4959]: I1007 13:49:01.747370 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkxdr" event={"ID":"a5295901-596b-447c-8d7a-bac77b140153","Type":"ContainerDied","Data":"8516b43df0c7fb6fe47ee62563ae339665cc2a77ba523d495be9955bf725b64b"} Oct 07 13:49:01 crc kubenswrapper[4959]: I1007 13:49:01.747814 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5brqg" podUID="11da31c7-46e5-4d9c-b11d-b67fae805f5d" containerName="registry-server" containerID="cri-o://ba8ee5e3680f017d6ab7b3d17688f06fb81b9f304f066d1cf0cde72c5539f6d9" gracePeriod=2 Oct 07 13:49:02 crc kubenswrapper[4959]: I1007 13:49:02.715861 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kkxdr" Oct 07 13:49:02 crc kubenswrapper[4959]: I1007 13:49:02.757452 4959 generic.go:334] "Generic (PLEG): container finished" podID="f0e27f2e-7704-4082-91e5-571830f218fc" containerID="7e53afb61245f016c470ea7180203eefef643a82a67cf7df3e2bfe4b44f7becd" exitCode=0 Oct 07 13:49:02 crc kubenswrapper[4959]: I1007 13:49:02.757515 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9bkvf" event={"ID":"f0e27f2e-7704-4082-91e5-571830f218fc","Type":"ContainerDied","Data":"7e53afb61245f016c470ea7180203eefef643a82a67cf7df3e2bfe4b44f7becd"} Oct 07 13:49:02 crc kubenswrapper[4959]: I1007 13:49:02.759683 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkxdr" event={"ID":"a5295901-596b-447c-8d7a-bac77b140153","Type":"ContainerDied","Data":"0f2e047cc089b406d7dd7577703ae16a1ffe2a89e2b898e814114fe5bc0601f9"} Oct 07 13:49:02 crc kubenswrapper[4959]: I1007 13:49:02.759724 4959 scope.go:117] "RemoveContainer" containerID="8516b43df0c7fb6fe47ee62563ae339665cc2a77ba523d495be9955bf725b64b" Oct 07 13:49:02 crc kubenswrapper[4959]: I1007 13:49:02.759746 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kkxdr" Oct 07 13:49:02 crc kubenswrapper[4959]: I1007 13:49:02.857738 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5295901-596b-447c-8d7a-bac77b140153-utilities\") pod \"a5295901-596b-447c-8d7a-bac77b140153\" (UID: \"a5295901-596b-447c-8d7a-bac77b140153\") " Oct 07 13:49:02 crc kubenswrapper[4959]: I1007 13:49:02.857860 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grzcm\" (UniqueName: \"kubernetes.io/projected/a5295901-596b-447c-8d7a-bac77b140153-kube-api-access-grzcm\") pod \"a5295901-596b-447c-8d7a-bac77b140153\" (UID: \"a5295901-596b-447c-8d7a-bac77b140153\") " Oct 07 13:49:02 crc kubenswrapper[4959]: I1007 13:49:02.857880 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5295901-596b-447c-8d7a-bac77b140153-catalog-content\") pod \"a5295901-596b-447c-8d7a-bac77b140153\" (UID: \"a5295901-596b-447c-8d7a-bac77b140153\") " Oct 07 13:49:02 crc kubenswrapper[4959]: I1007 13:49:02.858729 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5295901-596b-447c-8d7a-bac77b140153-utilities" (OuterVolumeSpecName: "utilities") pod "a5295901-596b-447c-8d7a-bac77b140153" (UID: "a5295901-596b-447c-8d7a-bac77b140153"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:49:02 crc kubenswrapper[4959]: I1007 13:49:02.865594 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5295901-596b-447c-8d7a-bac77b140153-kube-api-access-grzcm" (OuterVolumeSpecName: "kube-api-access-grzcm") pod "a5295901-596b-447c-8d7a-bac77b140153" (UID: "a5295901-596b-447c-8d7a-bac77b140153"). InnerVolumeSpecName "kube-api-access-grzcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:49:02 crc kubenswrapper[4959]: I1007 13:49:02.892423 4959 scope.go:117] "RemoveContainer" containerID="1fe1aa977e33c0e9309d7891c1abf7026211722775c574c75004397def0274f3" Oct 07 13:49:02 crc kubenswrapper[4959]: I1007 13:49:02.916789 4959 scope.go:117] "RemoveContainer" containerID="d310dd3b306a87bc16520f59be6687984d547c6d3eadc02cd9d9487dab392380" Oct 07 13:49:02 crc kubenswrapper[4959]: I1007 13:49:02.959556 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grzcm\" (UniqueName: \"kubernetes.io/projected/a5295901-596b-447c-8d7a-bac77b140153-kube-api-access-grzcm\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:02 crc kubenswrapper[4959]: I1007 13:49:02.959588 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5295901-596b-447c-8d7a-bac77b140153-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:03 crc kubenswrapper[4959]: I1007 13:49:03.185440 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9bkvf" Oct 07 13:49:03 crc kubenswrapper[4959]: I1007 13:49:03.262716 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0e27f2e-7704-4082-91e5-571830f218fc-utilities\") pod \"f0e27f2e-7704-4082-91e5-571830f218fc\" (UID: \"f0e27f2e-7704-4082-91e5-571830f218fc\") " Oct 07 13:49:03 crc kubenswrapper[4959]: I1007 13:49:03.263511 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0e27f2e-7704-4082-91e5-571830f218fc-utilities" (OuterVolumeSpecName: "utilities") pod "f0e27f2e-7704-4082-91e5-571830f218fc" (UID: "f0e27f2e-7704-4082-91e5-571830f218fc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:49:03 crc kubenswrapper[4959]: I1007 13:49:03.363539 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0e27f2e-7704-4082-91e5-571830f218fc-catalog-content\") pod \"f0e27f2e-7704-4082-91e5-571830f218fc\" (UID: \"f0e27f2e-7704-4082-91e5-571830f218fc\") " Oct 07 13:49:03 crc kubenswrapper[4959]: I1007 13:49:03.363743 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rw9ct\" (UniqueName: \"kubernetes.io/projected/f0e27f2e-7704-4082-91e5-571830f218fc-kube-api-access-rw9ct\") pod \"f0e27f2e-7704-4082-91e5-571830f218fc\" (UID: \"f0e27f2e-7704-4082-91e5-571830f218fc\") " Oct 07 13:49:03 crc kubenswrapper[4959]: I1007 13:49:03.364027 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0e27f2e-7704-4082-91e5-571830f218fc-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:03 crc kubenswrapper[4959]: I1007 13:49:03.370620 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0e27f2e-7704-4082-91e5-571830f218fc-kube-api-access-rw9ct" (OuterVolumeSpecName: "kube-api-access-rw9ct") pod "f0e27f2e-7704-4082-91e5-571830f218fc" (UID: "f0e27f2e-7704-4082-91e5-571830f218fc"). InnerVolumeSpecName "kube-api-access-rw9ct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:49:03 crc kubenswrapper[4959]: I1007 13:49:03.464832 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rw9ct\" (UniqueName: \"kubernetes.io/projected/f0e27f2e-7704-4082-91e5-571830f218fc-kube-api-access-rw9ct\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:03 crc kubenswrapper[4959]: I1007 13:49:03.766687 4959 generic.go:334] "Generic (PLEG): container finished" podID="11da31c7-46e5-4d9c-b11d-b67fae805f5d" containerID="ba8ee5e3680f017d6ab7b3d17688f06fb81b9f304f066d1cf0cde72c5539f6d9" exitCode=0 Oct 07 13:49:03 crc kubenswrapper[4959]: I1007 13:49:03.766758 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5brqg" event={"ID":"11da31c7-46e5-4d9c-b11d-b67fae805f5d","Type":"ContainerDied","Data":"ba8ee5e3680f017d6ab7b3d17688f06fb81b9f304f066d1cf0cde72c5539f6d9"} Oct 07 13:49:03 crc kubenswrapper[4959]: I1007 13:49:03.769938 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9bkvf" event={"ID":"f0e27f2e-7704-4082-91e5-571830f218fc","Type":"ContainerDied","Data":"7d08893e260f23e4dd96c0b2263c6acbc2f4f9a7be8b06b413570bf468519447"} Oct 07 13:49:03 crc kubenswrapper[4959]: I1007 13:49:03.769998 4959 scope.go:117] "RemoveContainer" containerID="7e53afb61245f016c470ea7180203eefef643a82a67cf7df3e2bfe4b44f7becd" Oct 07 13:49:03 crc kubenswrapper[4959]: I1007 13:49:03.770007 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9bkvf" Oct 07 13:49:03 crc kubenswrapper[4959]: I1007 13:49:03.782190 4959 scope.go:117] "RemoveContainer" containerID="a529d7542c01271a08ed640b28c836c3f83759c5ed3bf7f2e275b697003efa1d" Oct 07 13:49:03 crc kubenswrapper[4959]: I1007 13:49:03.805225 4959 scope.go:117] "RemoveContainer" containerID="df3f22c056ab913f07fe26d3d08be8852e1249beabfafa9ee7f36c96334a4ad9" Oct 07 13:49:04 crc kubenswrapper[4959]: I1007 13:49:04.069284 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pxlrd"] Oct 07 13:49:04 crc kubenswrapper[4959]: I1007 13:49:04.069484 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pxlrd" podUID="087e1130-9588-4d57-a552-6c868170c308" containerName="registry-server" containerID="cri-o://009ef1660013fd06e51f92374b03c12ffbdf04b5f91932b988e61bdb6b125d59" gracePeriod=2 Oct 07 13:49:05 crc kubenswrapper[4959]: I1007 13:49:05.115873 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5brqg" Oct 07 13:49:05 crc kubenswrapper[4959]: I1007 13:49:05.120374 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5295901-596b-447c-8d7a-bac77b140153-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a5295901-596b-447c-8d7a-bac77b140153" (UID: "a5295901-596b-447c-8d7a-bac77b140153"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:49:05 crc kubenswrapper[4959]: I1007 13:49:05.134063 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0e27f2e-7704-4082-91e5-571830f218fc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f0e27f2e-7704-4082-91e5-571830f218fc" (UID: "f0e27f2e-7704-4082-91e5-571830f218fc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:49:05 crc kubenswrapper[4959]: I1007 13:49:05.185997 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0e27f2e-7704-4082-91e5-571830f218fc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:05 crc kubenswrapper[4959]: I1007 13:49:05.186029 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5295901-596b-447c-8d7a-bac77b140153-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:05 crc kubenswrapper[4959]: I1007 13:49:05.189892 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kkxdr"] Oct 07 13:49:05 crc kubenswrapper[4959]: I1007 13:49:05.193291 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kkxdr"] Oct 07 13:49:05 crc kubenswrapper[4959]: I1007 13:49:05.287023 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11da31c7-46e5-4d9c-b11d-b67fae805f5d-catalog-content\") pod \"11da31c7-46e5-4d9c-b11d-b67fae805f5d\" (UID: \"11da31c7-46e5-4d9c-b11d-b67fae805f5d\") " Oct 07 13:49:05 crc kubenswrapper[4959]: I1007 13:49:05.287217 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11da31c7-46e5-4d9c-b11d-b67fae805f5d-utilities\") pod \"11da31c7-46e5-4d9c-b11d-b67fae805f5d\" (UID: \"11da31c7-46e5-4d9c-b11d-b67fae805f5d\") " Oct 07 13:49:05 crc kubenswrapper[4959]: I1007 13:49:05.287286 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk75g\" (UniqueName: \"kubernetes.io/projected/11da31c7-46e5-4d9c-b11d-b67fae805f5d-kube-api-access-tk75g\") pod \"11da31c7-46e5-4d9c-b11d-b67fae805f5d\" (UID: \"11da31c7-46e5-4d9c-b11d-b67fae805f5d\") " Oct 07 13:49:05 crc kubenswrapper[4959]: I1007 13:49:05.288890 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11da31c7-46e5-4d9c-b11d-b67fae805f5d-utilities" (OuterVolumeSpecName: "utilities") pod "11da31c7-46e5-4d9c-b11d-b67fae805f5d" (UID: "11da31c7-46e5-4d9c-b11d-b67fae805f5d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:49:05 crc kubenswrapper[4959]: I1007 13:49:05.291110 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11da31c7-46e5-4d9c-b11d-b67fae805f5d-kube-api-access-tk75g" (OuterVolumeSpecName: "kube-api-access-tk75g") pod "11da31c7-46e5-4d9c-b11d-b67fae805f5d" (UID: "11da31c7-46e5-4d9c-b11d-b67fae805f5d"). InnerVolumeSpecName "kube-api-access-tk75g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:49:05 crc kubenswrapper[4959]: I1007 13:49:05.305684 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9bkvf"] Oct 07 13:49:05 crc kubenswrapper[4959]: I1007 13:49:05.308060 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9bkvf"] Oct 07 13:49:05 crc kubenswrapper[4959]: I1007 13:49:05.311211 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11da31c7-46e5-4d9c-b11d-b67fae805f5d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "11da31c7-46e5-4d9c-b11d-b67fae805f5d" (UID: "11da31c7-46e5-4d9c-b11d-b67fae805f5d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:49:05 crc kubenswrapper[4959]: I1007 13:49:05.388763 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk75g\" (UniqueName: \"kubernetes.io/projected/11da31c7-46e5-4d9c-b11d-b67fae805f5d-kube-api-access-tk75g\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:05 crc kubenswrapper[4959]: I1007 13:49:05.388823 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11da31c7-46e5-4d9c-b11d-b67fae805f5d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:05 crc kubenswrapper[4959]: I1007 13:49:05.388836 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11da31c7-46e5-4d9c-b11d-b67fae805f5d-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:05 crc kubenswrapper[4959]: I1007 13:49:05.784810 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5brqg" event={"ID":"11da31c7-46e5-4d9c-b11d-b67fae805f5d","Type":"ContainerDied","Data":"d36ad2c834f36017d2e03645259f49c7465c493686d70a19d0bf9060e7a03ce1"} Oct 07 13:49:05 crc kubenswrapper[4959]: I1007 13:49:05.784868 4959 scope.go:117] "RemoveContainer" containerID="ba8ee5e3680f017d6ab7b3d17688f06fb81b9f304f066d1cf0cde72c5539f6d9" Oct 07 13:49:05 crc kubenswrapper[4959]: I1007 13:49:05.784935 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5brqg" Oct 07 13:49:05 crc kubenswrapper[4959]: I1007 13:49:05.801626 4959 scope.go:117] "RemoveContainer" containerID="1321ae2985977e169b34124567d766faa3d33170d43f74a6d9f0cec83dd81130" Oct 07 13:49:05 crc kubenswrapper[4959]: I1007 13:49:05.823510 4959 scope.go:117] "RemoveContainer" containerID="bdefccd76c25c5312021d53ad897f45a013c7ba5bb4e8ece974659cc95c0b843" Oct 07 13:49:05 crc kubenswrapper[4959]: I1007 13:49:05.825279 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5brqg"] Oct 07 13:49:05 crc kubenswrapper[4959]: I1007 13:49:05.828185 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5brqg"] Oct 07 13:49:06 crc kubenswrapper[4959]: I1007 13:49:06.661321 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11da31c7-46e5-4d9c-b11d-b67fae805f5d" path="/var/lib/kubelet/pods/11da31c7-46e5-4d9c-b11d-b67fae805f5d/volumes" Oct 07 13:49:06 crc kubenswrapper[4959]: I1007 13:49:06.662226 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5295901-596b-447c-8d7a-bac77b140153" path="/var/lib/kubelet/pods/a5295901-596b-447c-8d7a-bac77b140153/volumes" Oct 07 13:49:06 crc kubenswrapper[4959]: I1007 13:49:06.662999 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0e27f2e-7704-4082-91e5-571830f218fc" path="/var/lib/kubelet/pods/f0e27f2e-7704-4082-91e5-571830f218fc/volumes" Oct 07 13:49:06 crc kubenswrapper[4959]: I1007 13:49:06.790891 4959 generic.go:334] "Generic (PLEG): container finished" podID="087e1130-9588-4d57-a552-6c868170c308" containerID="009ef1660013fd06e51f92374b03c12ffbdf04b5f91932b988e61bdb6b125d59" exitCode=0 Oct 07 13:49:06 crc kubenswrapper[4959]: I1007 13:49:06.790966 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxlrd" event={"ID":"087e1130-9588-4d57-a552-6c868170c308","Type":"ContainerDied","Data":"009ef1660013fd06e51f92374b03c12ffbdf04b5f91932b988e61bdb6b125d59"} Oct 07 13:49:09 crc kubenswrapper[4959]: I1007 13:49:09.119538 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pxlrd" Oct 07 13:49:09 crc kubenswrapper[4959]: I1007 13:49:09.235379 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/087e1130-9588-4d57-a552-6c868170c308-utilities\") pod \"087e1130-9588-4d57-a552-6c868170c308\" (UID: \"087e1130-9588-4d57-a552-6c868170c308\") " Oct 07 13:49:09 crc kubenswrapper[4959]: I1007 13:49:09.235514 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/087e1130-9588-4d57-a552-6c868170c308-catalog-content\") pod \"087e1130-9588-4d57-a552-6c868170c308\" (UID: \"087e1130-9588-4d57-a552-6c868170c308\") " Oct 07 13:49:09 crc kubenswrapper[4959]: I1007 13:49:09.235578 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sq8jg\" (UniqueName: \"kubernetes.io/projected/087e1130-9588-4d57-a552-6c868170c308-kube-api-access-sq8jg\") pod \"087e1130-9588-4d57-a552-6c868170c308\" (UID: \"087e1130-9588-4d57-a552-6c868170c308\") " Oct 07 13:49:09 crc kubenswrapper[4959]: I1007 13:49:09.236152 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/087e1130-9588-4d57-a552-6c868170c308-utilities" (OuterVolumeSpecName: "utilities") pod "087e1130-9588-4d57-a552-6c868170c308" (UID: "087e1130-9588-4d57-a552-6c868170c308"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:49:09 crc kubenswrapper[4959]: I1007 13:49:09.245409 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/087e1130-9588-4d57-a552-6c868170c308-kube-api-access-sq8jg" (OuterVolumeSpecName: "kube-api-access-sq8jg") pod "087e1130-9588-4d57-a552-6c868170c308" (UID: "087e1130-9588-4d57-a552-6c868170c308"). InnerVolumeSpecName "kube-api-access-sq8jg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:49:09 crc kubenswrapper[4959]: I1007 13:49:09.310502 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/087e1130-9588-4d57-a552-6c868170c308-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "087e1130-9588-4d57-a552-6c868170c308" (UID: "087e1130-9588-4d57-a552-6c868170c308"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:49:09 crc kubenswrapper[4959]: I1007 13:49:09.336545 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/087e1130-9588-4d57-a552-6c868170c308-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:09 crc kubenswrapper[4959]: I1007 13:49:09.336567 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/087e1130-9588-4d57-a552-6c868170c308-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:09 crc kubenswrapper[4959]: I1007 13:49:09.336577 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sq8jg\" (UniqueName: \"kubernetes.io/projected/087e1130-9588-4d57-a552-6c868170c308-kube-api-access-sq8jg\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:09 crc kubenswrapper[4959]: I1007 13:49:09.808044 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxlrd" event={"ID":"087e1130-9588-4d57-a552-6c868170c308","Type":"ContainerDied","Data":"d71a93a1d58c858fbda701797685b73a253cae5d7ee4d04d61c6c092b368fe87"} Oct 07 13:49:09 crc kubenswrapper[4959]: I1007 13:49:09.808087 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pxlrd" Oct 07 13:49:09 crc kubenswrapper[4959]: I1007 13:49:09.808095 4959 scope.go:117] "RemoveContainer" containerID="009ef1660013fd06e51f92374b03c12ffbdf04b5f91932b988e61bdb6b125d59" Oct 07 13:49:09 crc kubenswrapper[4959]: I1007 13:49:09.823402 4959 scope.go:117] "RemoveContainer" containerID="f87747fb36379f2e7835619556bc4e51bc21be04b9349d85ffdb200b8020ffd2" Oct 07 13:49:09 crc kubenswrapper[4959]: I1007 13:49:09.834655 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pxlrd"] Oct 07 13:49:09 crc kubenswrapper[4959]: I1007 13:49:09.838212 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pxlrd"] Oct 07 13:49:09 crc kubenswrapper[4959]: I1007 13:49:09.865321 4959 scope.go:117] "RemoveContainer" containerID="04e934f377283b5ed1b5e8feacc637ff2e4499fa1317874e293f19b9d4dc3651" Oct 07 13:49:10 crc kubenswrapper[4959]: I1007 13:49:10.660215 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="087e1130-9588-4d57-a552-6c868170c308" path="/var/lib/kubelet/pods/087e1130-9588-4d57-a552-6c868170c308/volumes" Oct 07 13:49:19 crc kubenswrapper[4959]: I1007 13:49:19.744762 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" podUID="ced32e5c-2931-4a96-a779-1c1c6b71243e" containerName="oauth-openshift" containerID="cri-o://91579ad9e4cf546e01320e6a4a2badc05b6ad3da8ae5dde2d16d52a509cbfdfe" gracePeriod=15 Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.170180 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.195633 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-router-certs\") pod \"ced32e5c-2931-4a96-a779-1c1c6b71243e\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.195751 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-user-template-error\") pod \"ced32e5c-2931-4a96-a779-1c1c6b71243e\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.195797 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ced32e5c-2931-4a96-a779-1c1c6b71243e-audit-dir\") pod \"ced32e5c-2931-4a96-a779-1c1c6b71243e\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.195873 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-cliconfig\") pod \"ced32e5c-2931-4a96-a779-1c1c6b71243e\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.195911 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-service-ca\") pod \"ced32e5c-2931-4a96-a779-1c1c6b71243e\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.195952 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-serving-cert\") pod \"ced32e5c-2931-4a96-a779-1c1c6b71243e\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.195997 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-ocp-branding-template\") pod \"ced32e5c-2931-4a96-a779-1c1c6b71243e\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.196036 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-user-template-provider-selection\") pod \"ced32e5c-2931-4a96-a779-1c1c6b71243e\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.196079 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ced32e5c-2931-4a96-a779-1c1c6b71243e-audit-policies\") pod \"ced32e5c-2931-4a96-a779-1c1c6b71243e\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.196130 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-user-idp-0-file-data\") pod \"ced32e5c-2931-4a96-a779-1c1c6b71243e\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.196166 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-user-template-login\") pod \"ced32e5c-2931-4a96-a779-1c1c6b71243e\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.196199 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-trusted-ca-bundle\") pod \"ced32e5c-2931-4a96-a779-1c1c6b71243e\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.196235 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpw6q\" (UniqueName: \"kubernetes.io/projected/ced32e5c-2931-4a96-a779-1c1c6b71243e-kube-api-access-tpw6q\") pod \"ced32e5c-2931-4a96-a779-1c1c6b71243e\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.196269 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-session\") pod \"ced32e5c-2931-4a96-a779-1c1c6b71243e\" (UID: \"ced32e5c-2931-4a96-a779-1c1c6b71243e\") " Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.198310 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "ced32e5c-2931-4a96-a779-1c1c6b71243e" (UID: "ced32e5c-2931-4a96-a779-1c1c6b71243e"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.217211 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ced32e5c-2931-4a96-a779-1c1c6b71243e-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "ced32e5c-2931-4a96-a779-1c1c6b71243e" (UID: "ced32e5c-2931-4a96-a779-1c1c6b71243e"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.218553 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ced32e5c-2931-4a96-a779-1c1c6b71243e-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "ced32e5c-2931-4a96-a779-1c1c6b71243e" (UID: "ced32e5c-2931-4a96-a779-1c1c6b71243e"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.218967 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "ced32e5c-2931-4a96-a779-1c1c6b71243e" (UID: "ced32e5c-2931-4a96-a779-1c1c6b71243e"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.219386 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "ced32e5c-2931-4a96-a779-1c1c6b71243e" (UID: "ced32e5c-2931-4a96-a779-1c1c6b71243e"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.229681 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "ced32e5c-2931-4a96-a779-1c1c6b71243e" (UID: "ced32e5c-2931-4a96-a779-1c1c6b71243e"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.245319 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "ced32e5c-2931-4a96-a779-1c1c6b71243e" (UID: "ced32e5c-2931-4a96-a779-1c1c6b71243e"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.264993 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ced32e5c-2931-4a96-a779-1c1c6b71243e-kube-api-access-tpw6q" (OuterVolumeSpecName: "kube-api-access-tpw6q") pod "ced32e5c-2931-4a96-a779-1c1c6b71243e" (UID: "ced32e5c-2931-4a96-a779-1c1c6b71243e"). InnerVolumeSpecName "kube-api-access-tpw6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.267572 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "ced32e5c-2931-4a96-a779-1c1c6b71243e" (UID: "ced32e5c-2931-4a96-a779-1c1c6b71243e"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.271852 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "ced32e5c-2931-4a96-a779-1c1c6b71243e" (UID: "ced32e5c-2931-4a96-a779-1c1c6b71243e"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.272038 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "ced32e5c-2931-4a96-a779-1c1c6b71243e" (UID: "ced32e5c-2931-4a96-a779-1c1c6b71243e"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.272330 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "ced32e5c-2931-4a96-a779-1c1c6b71243e" (UID: "ced32e5c-2931-4a96-a779-1c1c6b71243e"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.272519 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "ced32e5c-2931-4a96-a779-1c1c6b71243e" (UID: "ced32e5c-2931-4a96-a779-1c1c6b71243e"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.272546 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "ced32e5c-2931-4a96-a779-1c1c6b71243e" (UID: "ced32e5c-2931-4a96-a779-1c1c6b71243e"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.298753 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.298787 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.298799 4959 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ced32e5c-2931-4a96-a779-1c1c6b71243e-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.298814 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.298823 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.298833 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.298842 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.298851 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.298861 4959 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ced32e5c-2931-4a96-a779-1c1c6b71243e-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.298870 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.298878 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.298886 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.298896 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpw6q\" (UniqueName: \"kubernetes.io/projected/ced32e5c-2931-4a96-a779-1c1c6b71243e-kube-api-access-tpw6q\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.298904 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ced32e5c-2931-4a96-a779-1c1c6b71243e-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.874036 4959 generic.go:334] "Generic (PLEG): container finished" podID="ced32e5c-2931-4a96-a779-1c1c6b71243e" containerID="91579ad9e4cf546e01320e6a4a2badc05b6ad3da8ae5dde2d16d52a509cbfdfe" exitCode=0 Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.874089 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" event={"ID":"ced32e5c-2931-4a96-a779-1c1c6b71243e","Type":"ContainerDied","Data":"91579ad9e4cf546e01320e6a4a2badc05b6ad3da8ae5dde2d16d52a509cbfdfe"} Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.874159 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" event={"ID":"ced32e5c-2931-4a96-a779-1c1c6b71243e","Type":"ContainerDied","Data":"69b3d8090a7773777e3906944bb7694b711e47ef8f695739356b03ee05af5146"} Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.874176 4959 scope.go:117] "RemoveContainer" containerID="91579ad9e4cf546e01320e6a4a2badc05b6ad3da8ae5dde2d16d52a509cbfdfe" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.874233 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-67jl7" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.908449 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-67jl7"] Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.911175 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-67jl7"] Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.913646 4959 scope.go:117] "RemoveContainer" containerID="91579ad9e4cf546e01320e6a4a2badc05b6ad3da8ae5dde2d16d52a509cbfdfe" Oct 07 13:49:20 crc kubenswrapper[4959]: E1007 13:49:20.914220 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91579ad9e4cf546e01320e6a4a2badc05b6ad3da8ae5dde2d16d52a509cbfdfe\": container with ID starting with 91579ad9e4cf546e01320e6a4a2badc05b6ad3da8ae5dde2d16d52a509cbfdfe not found: ID does not exist" containerID="91579ad9e4cf546e01320e6a4a2badc05b6ad3da8ae5dde2d16d52a509cbfdfe" Oct 07 13:49:20 crc kubenswrapper[4959]: I1007 13:49:20.914272 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91579ad9e4cf546e01320e6a4a2badc05b6ad3da8ae5dde2d16d52a509cbfdfe"} err="failed to get container status \"91579ad9e4cf546e01320e6a4a2badc05b6ad3da8ae5dde2d16d52a509cbfdfe\": rpc error: code = NotFound desc = could not find container \"91579ad9e4cf546e01320e6a4a2badc05b6ad3da8ae5dde2d16d52a509cbfdfe\": container with ID starting with 91579ad9e4cf546e01320e6a4a2badc05b6ad3da8ae5dde2d16d52a509cbfdfe not found: ID does not exist" Oct 07 13:49:22 crc kubenswrapper[4959]: I1007 13:49:22.659964 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ced32e5c-2931-4a96-a779-1c1c6b71243e" path="/var/lib/kubelet/pods/ced32e5c-2931-4a96-a779-1c1c6b71243e/volumes" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.978770 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-fdd74686d-mmzhb"] Oct 07 13:49:26 crc kubenswrapper[4959]: E1007 13:49:26.979437 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="087e1130-9588-4d57-a552-6c868170c308" containerName="extract-utilities" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.979458 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="087e1130-9588-4d57-a552-6c868170c308" containerName="extract-utilities" Oct 07 13:49:26 crc kubenswrapper[4959]: E1007 13:49:26.979482 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11da31c7-46e5-4d9c-b11d-b67fae805f5d" containerName="extract-utilities" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.979495 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="11da31c7-46e5-4d9c-b11d-b67fae805f5d" containerName="extract-utilities" Oct 07 13:49:26 crc kubenswrapper[4959]: E1007 13:49:26.979508 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5295901-596b-447c-8d7a-bac77b140153" containerName="registry-server" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.979522 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5295901-596b-447c-8d7a-bac77b140153" containerName="registry-server" Oct 07 13:49:26 crc kubenswrapper[4959]: E1007 13:49:26.979543 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5295901-596b-447c-8d7a-bac77b140153" containerName="extract-content" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.979556 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5295901-596b-447c-8d7a-bac77b140153" containerName="extract-content" Oct 07 13:49:26 crc kubenswrapper[4959]: E1007 13:49:26.979578 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="087e1130-9588-4d57-a552-6c868170c308" containerName="extract-content" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.979589 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="087e1130-9588-4d57-a552-6c868170c308" containerName="extract-content" Oct 07 13:49:26 crc kubenswrapper[4959]: E1007 13:49:26.979607 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="087e1130-9588-4d57-a552-6c868170c308" containerName="registry-server" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.979620 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="087e1130-9588-4d57-a552-6c868170c308" containerName="registry-server" Oct 07 13:49:26 crc kubenswrapper[4959]: E1007 13:49:26.979634 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5295901-596b-447c-8d7a-bac77b140153" containerName="extract-utilities" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.979647 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5295901-596b-447c-8d7a-bac77b140153" containerName="extract-utilities" Oct 07 13:49:26 crc kubenswrapper[4959]: E1007 13:49:26.979661 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="601096a4-0f0e-4c06-bce3-bdccdd1e370b" containerName="pruner" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.979673 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="601096a4-0f0e-4c06-bce3-bdccdd1e370b" containerName="pruner" Oct 07 13:49:26 crc kubenswrapper[4959]: E1007 13:49:26.979688 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0e27f2e-7704-4082-91e5-571830f218fc" containerName="extract-utilities" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.979702 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0e27f2e-7704-4082-91e5-571830f218fc" containerName="extract-utilities" Oct 07 13:49:26 crc kubenswrapper[4959]: E1007 13:49:26.979720 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50a8e4e2-c0eb-4fba-ac68-03824eb1f99f" containerName="pruner" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.979731 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="50a8e4e2-c0eb-4fba-ac68-03824eb1f99f" containerName="pruner" Oct 07 13:49:26 crc kubenswrapper[4959]: E1007 13:49:26.979748 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ced32e5c-2931-4a96-a779-1c1c6b71243e" containerName="oauth-openshift" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.979760 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ced32e5c-2931-4a96-a779-1c1c6b71243e" containerName="oauth-openshift" Oct 07 13:49:26 crc kubenswrapper[4959]: E1007 13:49:26.979782 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11da31c7-46e5-4d9c-b11d-b67fae805f5d" containerName="registry-server" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.979794 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="11da31c7-46e5-4d9c-b11d-b67fae805f5d" containerName="registry-server" Oct 07 13:49:26 crc kubenswrapper[4959]: E1007 13:49:26.979813 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0e27f2e-7704-4082-91e5-571830f218fc" containerName="registry-server" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.979824 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0e27f2e-7704-4082-91e5-571830f218fc" containerName="registry-server" Oct 07 13:49:26 crc kubenswrapper[4959]: E1007 13:49:26.979841 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0e27f2e-7704-4082-91e5-571830f218fc" containerName="extract-content" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.979852 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0e27f2e-7704-4082-91e5-571830f218fc" containerName="extract-content" Oct 07 13:49:26 crc kubenswrapper[4959]: E1007 13:49:26.979864 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11da31c7-46e5-4d9c-b11d-b67fae805f5d" containerName="extract-content" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.979876 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="11da31c7-46e5-4d9c-b11d-b67fae805f5d" containerName="extract-content" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.980018 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="50a8e4e2-c0eb-4fba-ac68-03824eb1f99f" containerName="pruner" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.980037 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="601096a4-0f0e-4c06-bce3-bdccdd1e370b" containerName="pruner" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.980054 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="087e1130-9588-4d57-a552-6c868170c308" containerName="registry-server" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.980072 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="11da31c7-46e5-4d9c-b11d-b67fae805f5d" containerName="registry-server" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.980092 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="ced32e5c-2931-4a96-a779-1c1c6b71243e" containerName="oauth-openshift" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.980135 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5295901-596b-447c-8d7a-bac77b140153" containerName="registry-server" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.980155 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0e27f2e-7704-4082-91e5-571830f218fc" containerName="registry-server" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.980749 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.983322 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.983526 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.983327 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.983893 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.984664 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.984715 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.985149 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.985180 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.987766 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.994378 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.994727 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 07 13:49:26 crc kubenswrapper[4959]: I1007 13:49:26.994929 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.002500 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.007906 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.010962 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.024438 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-fdd74686d-mmzhb"] Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.083096 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-user-template-login\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.083169 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.083214 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.083238 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-audit-policies\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.083261 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-user-template-error\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.083286 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.083311 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-audit-dir\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.083339 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-system-router-certs\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.083626 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.083671 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhc5t\" (UniqueName: \"kubernetes.io/projected/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-kube-api-access-vhc5t\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.083699 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-system-session\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.083801 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-system-service-ca\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.083833 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.083858 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.184911 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-system-service-ca\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.184949 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.184968 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.184990 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-user-template-login\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.185010 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.185038 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.185060 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-audit-policies\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.185748 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-user-template-error\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.185831 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.185885 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-audit-dir\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.185949 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-system-router-certs\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.186045 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-system-session\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.186044 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-audit-dir\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.186080 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.186161 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhc5t\" (UniqueName: \"kubernetes.io/projected/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-kube-api-access-vhc5t\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.186595 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.186942 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.186967 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-audit-policies\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.187367 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-system-service-ca\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.191577 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.191699 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.191807 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-system-router-certs\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.191959 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.202505 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-system-session\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.202548 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-user-template-login\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.203332 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-user-template-error\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.203435 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.206942 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhc5t\" (UniqueName: \"kubernetes.io/projected/54d5a9a1-4b6b-4aed-816d-d9f645f3732c-kube-api-access-vhc5t\") pod \"oauth-openshift-fdd74686d-mmzhb\" (UID: \"54d5a9a1-4b6b-4aed-816d-d9f645f3732c\") " pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.321903 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.611092 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-fdd74686d-mmzhb"] Oct 07 13:49:27 crc kubenswrapper[4959]: W1007 13:49:27.625317 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54d5a9a1_4b6b_4aed_816d_d9f645f3732c.slice/crio-35fe9fb56b7dc77deeb622a15cc97b7d7932896efa3d2af09fcf7c6e017ed63f WatchSource:0}: Error finding container 35fe9fb56b7dc77deeb622a15cc97b7d7932896efa3d2af09fcf7c6e017ed63f: Status 404 returned error can't find the container with id 35fe9fb56b7dc77deeb622a15cc97b7d7932896efa3d2af09fcf7c6e017ed63f Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.918129 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" event={"ID":"54d5a9a1-4b6b-4aed-816d-d9f645f3732c","Type":"ContainerStarted","Data":"a376a029bf841498cc51ce0e2c79cb10cdcd92dd6156a4435baffe485af2551f"} Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.918501 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.918519 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" event={"ID":"54d5a9a1-4b6b-4aed-816d-d9f645f3732c","Type":"ContainerStarted","Data":"35fe9fb56b7dc77deeb622a15cc97b7d7932896efa3d2af09fcf7c6e017ed63f"} Oct 07 13:49:27 crc kubenswrapper[4959]: I1007 13:49:27.937291 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" podStartSLOduration=33.937253085 podStartE2EDuration="33.937253085s" podCreationTimestamp="2025-10-07 13:48:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:49:27.934849338 +0000 UTC m=+250.018253683" watchObservedRunningTime="2025-10-07 13:49:27.937253085 +0000 UTC m=+250.020657410" Oct 07 13:49:28 crc kubenswrapper[4959]: I1007 13:49:28.223935 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-fdd74686d-mmzhb" Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.395556 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5r424"] Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.396174 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5r424" podUID="4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac" containerName="registry-server" containerID="cri-o://44ca7c831bca81e37b93e090d7dd71055c8d6605796726e271b35d43b47d707a" gracePeriod=30 Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.404483 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t7mrz"] Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.404670 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-t7mrz" podUID="ebb5441f-839a-408b-bd51-3460006c3d83" containerName="registry-server" containerID="cri-o://b2384da6a2e7f871634616e179a6b68fbbec8011b37cc0931b7babc6c27c9546" gracePeriod=30 Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.415900 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xwlzm"] Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.416081 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-xwlzm" podUID="cb8de785-a298-414c-998c-7bd0585966e1" containerName="marketplace-operator" containerID="cri-o://0127c5ab88c1d7a1993c76d818a43a226a5ee407b5f5aaa3ff686cc1d404aa63" gracePeriod=30 Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.423180 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c94wx"] Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.423407 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-c94wx" podUID="44b4f9d4-243a-46a8-ab8a-4fb5ae11816b" containerName="registry-server" containerID="cri-o://2948cb9150d3ce5ef0e539f41a521dbe6e02828b34f4ea21d9bb12593c8e4f6f" gracePeriod=30 Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.438840 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d6vp9"] Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.439093 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-d6vp9" podUID="339283f2-dc57-4316-a334-6d79fbf2be9b" containerName="registry-server" containerID="cri-o://2daa6f510281017ba2c87047c2b8f0b82762823189e3d76885a628912d0e67a0" gracePeriod=30 Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.441513 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-599kq"] Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.442690 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-599kq" Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.455912 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-599kq"] Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.523692 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgtxv\" (UniqueName: \"kubernetes.io/projected/922b72b0-51d8-4914-b740-5be4725ab083-kube-api-access-bgtxv\") pod \"marketplace-operator-79b997595-599kq\" (UID: \"922b72b0-51d8-4914-b740-5be4725ab083\") " pod="openshift-marketplace/marketplace-operator-79b997595-599kq" Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.523746 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/922b72b0-51d8-4914-b740-5be4725ab083-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-599kq\" (UID: \"922b72b0-51d8-4914-b740-5be4725ab083\") " pod="openshift-marketplace/marketplace-operator-79b997595-599kq" Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.523789 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/922b72b0-51d8-4914-b740-5be4725ab083-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-599kq\" (UID: \"922b72b0-51d8-4914-b740-5be4725ab083\") " pod="openshift-marketplace/marketplace-operator-79b997595-599kq" Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.624434 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/922b72b0-51d8-4914-b740-5be4725ab083-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-599kq\" (UID: \"922b72b0-51d8-4914-b740-5be4725ab083\") " pod="openshift-marketplace/marketplace-operator-79b997595-599kq" Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.624730 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/922b72b0-51d8-4914-b740-5be4725ab083-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-599kq\" (UID: \"922b72b0-51d8-4914-b740-5be4725ab083\") " pod="openshift-marketplace/marketplace-operator-79b997595-599kq" Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.624783 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgtxv\" (UniqueName: \"kubernetes.io/projected/922b72b0-51d8-4914-b740-5be4725ab083-kube-api-access-bgtxv\") pod \"marketplace-operator-79b997595-599kq\" (UID: \"922b72b0-51d8-4914-b740-5be4725ab083\") " pod="openshift-marketplace/marketplace-operator-79b997595-599kq" Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.626376 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/922b72b0-51d8-4914-b740-5be4725ab083-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-599kq\" (UID: \"922b72b0-51d8-4914-b740-5be4725ab083\") " pod="openshift-marketplace/marketplace-operator-79b997595-599kq" Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.630673 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/922b72b0-51d8-4914-b740-5be4725ab083-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-599kq\" (UID: \"922b72b0-51d8-4914-b740-5be4725ab083\") " pod="openshift-marketplace/marketplace-operator-79b997595-599kq" Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.652228 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgtxv\" (UniqueName: \"kubernetes.io/projected/922b72b0-51d8-4914-b740-5be4725ab083-kube-api-access-bgtxv\") pod \"marketplace-operator-79b997595-599kq\" (UID: \"922b72b0-51d8-4914-b740-5be4725ab083\") " pod="openshift-marketplace/marketplace-operator-79b997595-599kq" Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.786134 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-599kq" Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.844329 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5r424" Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.929079 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhdvl\" (UniqueName: \"kubernetes.io/projected/4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac-kube-api-access-qhdvl\") pod \"4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac\" (UID: \"4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac\") " Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.929386 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac-utilities\") pod \"4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac\" (UID: \"4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac\") " Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.929490 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac-catalog-content\") pod \"4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac\" (UID: \"4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac\") " Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.930356 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac-utilities" (OuterVolumeSpecName: "utilities") pod "4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac" (UID: "4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.938529 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac-kube-api-access-qhdvl" (OuterVolumeSpecName: "kube-api-access-qhdvl") pod "4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac" (UID: "4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac"). InnerVolumeSpecName "kube-api-access-qhdvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.971428 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t7mrz" Oct 07 13:49:44 crc kubenswrapper[4959]: I1007 13:49:44.987199 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac" (UID: "4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.014569 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d6vp9" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.019069 4959 generic.go:334] "Generic (PLEG): container finished" podID="339283f2-dc57-4316-a334-6d79fbf2be9b" containerID="2daa6f510281017ba2c87047c2b8f0b82762823189e3d76885a628912d0e67a0" exitCode=0 Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.019141 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d6vp9" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.019136 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d6vp9" event={"ID":"339283f2-dc57-4316-a334-6d79fbf2be9b","Type":"ContainerDied","Data":"2daa6f510281017ba2c87047c2b8f0b82762823189e3d76885a628912d0e67a0"} Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.019236 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d6vp9" event={"ID":"339283f2-dc57-4316-a334-6d79fbf2be9b","Type":"ContainerDied","Data":"361cf57d8a72a237d4bab5b98cd68a869d83fcef7329840871147fdac46e2121"} Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.019293 4959 scope.go:117] "RemoveContainer" containerID="2daa6f510281017ba2c87047c2b8f0b82762823189e3d76885a628912d0e67a0" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.021681 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xwlzm" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.023395 4959 generic.go:334] "Generic (PLEG): container finished" podID="ebb5441f-839a-408b-bd51-3460006c3d83" containerID="b2384da6a2e7f871634616e179a6b68fbbec8011b37cc0931b7babc6c27c9546" exitCode=0 Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.023558 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7mrz" event={"ID":"ebb5441f-839a-408b-bd51-3460006c3d83","Type":"ContainerDied","Data":"b2384da6a2e7f871634616e179a6b68fbbec8011b37cc0931b7babc6c27c9546"} Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.023614 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7mrz" event={"ID":"ebb5441f-839a-408b-bd51-3460006c3d83","Type":"ContainerDied","Data":"07fad97d5e8a0ee517b2c0b72e6dbe1b1f2a868e708e11455ab0cf1da003c6a9"} Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.023750 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t7mrz" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.030731 4959 generic.go:334] "Generic (PLEG): container finished" podID="44b4f9d4-243a-46a8-ab8a-4fb5ae11816b" containerID="2948cb9150d3ce5ef0e539f41a521dbe6e02828b34f4ea21d9bb12593c8e4f6f" exitCode=0 Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.030844 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c94wx" event={"ID":"44b4f9d4-243a-46a8-ab8a-4fb5ae11816b","Type":"ContainerDied","Data":"2948cb9150d3ce5ef0e539f41a521dbe6e02828b34f4ea21d9bb12593c8e4f6f"} Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.032260 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebb5441f-839a-408b-bd51-3460006c3d83-catalog-content\") pod \"ebb5441f-839a-408b-bd51-3460006c3d83\" (UID: \"ebb5441f-839a-408b-bd51-3460006c3d83\") " Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.032384 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zw8q2\" (UniqueName: \"kubernetes.io/projected/339283f2-dc57-4316-a334-6d79fbf2be9b-kube-api-access-zw8q2\") pod \"339283f2-dc57-4316-a334-6d79fbf2be9b\" (UID: \"339283f2-dc57-4316-a334-6d79fbf2be9b\") " Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.032867 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/339283f2-dc57-4316-a334-6d79fbf2be9b-catalog-content\") pod \"339283f2-dc57-4316-a334-6d79fbf2be9b\" (UID: \"339283f2-dc57-4316-a334-6d79fbf2be9b\") " Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.032908 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebb5441f-839a-408b-bd51-3460006c3d83-utilities\") pod \"ebb5441f-839a-408b-bd51-3460006c3d83\" (UID: \"ebb5441f-839a-408b-bd51-3460006c3d83\") " Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.032942 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/339283f2-dc57-4316-a334-6d79fbf2be9b-utilities\") pod \"339283f2-dc57-4316-a334-6d79fbf2be9b\" (UID: \"339283f2-dc57-4316-a334-6d79fbf2be9b\") " Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.032980 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vvdz\" (UniqueName: \"kubernetes.io/projected/ebb5441f-839a-408b-bd51-3460006c3d83-kube-api-access-7vvdz\") pod \"ebb5441f-839a-408b-bd51-3460006c3d83\" (UID: \"ebb5441f-839a-408b-bd51-3460006c3d83\") " Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.033013 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cb8de785-a298-414c-998c-7bd0585966e1-marketplace-trusted-ca\") pod \"cb8de785-a298-414c-998c-7bd0585966e1\" (UID: \"cb8de785-a298-414c-998c-7bd0585966e1\") " Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.033731 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6d6vt\" (UniqueName: \"kubernetes.io/projected/cb8de785-a298-414c-998c-7bd0585966e1-kube-api-access-6d6vt\") pod \"cb8de785-a298-414c-998c-7bd0585966e1\" (UID: \"cb8de785-a298-414c-998c-7bd0585966e1\") " Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.033852 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/cb8de785-a298-414c-998c-7bd0585966e1-marketplace-operator-metrics\") pod \"cb8de785-a298-414c-998c-7bd0585966e1\" (UID: \"cb8de785-a298-414c-998c-7bd0585966e1\") " Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.034348 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhdvl\" (UniqueName: \"kubernetes.io/projected/4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac-kube-api-access-qhdvl\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.034373 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.034389 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.034710 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/339283f2-dc57-4316-a334-6d79fbf2be9b-utilities" (OuterVolumeSpecName: "utilities") pod "339283f2-dc57-4316-a334-6d79fbf2be9b" (UID: "339283f2-dc57-4316-a334-6d79fbf2be9b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.034822 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebb5441f-839a-408b-bd51-3460006c3d83-utilities" (OuterVolumeSpecName: "utilities") pod "ebb5441f-839a-408b-bd51-3460006c3d83" (UID: "ebb5441f-839a-408b-bd51-3460006c3d83"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.035450 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb8de785-a298-414c-998c-7bd0585966e1-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "cb8de785-a298-414c-998c-7bd0585966e1" (UID: "cb8de785-a298-414c-998c-7bd0585966e1"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.035547 4959 generic.go:334] "Generic (PLEG): container finished" podID="4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac" containerID="44ca7c831bca81e37b93e090d7dd71055c8d6605796726e271b35d43b47d707a" exitCode=0 Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.035665 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5r424" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.036174 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5r424" event={"ID":"4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac","Type":"ContainerDied","Data":"44ca7c831bca81e37b93e090d7dd71055c8d6605796726e271b35d43b47d707a"} Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.036197 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5r424" event={"ID":"4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac","Type":"ContainerDied","Data":"a6359874f8f2aa856f62bb09b77413cf993b7de19e61e8b2144bdc2ea74c4a08"} Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.043550 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb8de785-a298-414c-998c-7bd0585966e1-kube-api-access-6d6vt" (OuterVolumeSpecName: "kube-api-access-6d6vt") pod "cb8de785-a298-414c-998c-7bd0585966e1" (UID: "cb8de785-a298-414c-998c-7bd0585966e1"). InnerVolumeSpecName "kube-api-access-6d6vt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.058476 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb8de785-a298-414c-998c-7bd0585966e1-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "cb8de785-a298-414c-998c-7bd0585966e1" (UID: "cb8de785-a298-414c-998c-7bd0585966e1"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.061923 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/339283f2-dc57-4316-a334-6d79fbf2be9b-kube-api-access-zw8q2" (OuterVolumeSpecName: "kube-api-access-zw8q2") pod "339283f2-dc57-4316-a334-6d79fbf2be9b" (UID: "339283f2-dc57-4316-a334-6d79fbf2be9b"). InnerVolumeSpecName "kube-api-access-zw8q2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.062592 4959 scope.go:117] "RemoveContainer" containerID="00a246533cb95ee3510722163a228aba26a544d17b441a6cb7e496d9ea99c053" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.062474 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebb5441f-839a-408b-bd51-3460006c3d83-kube-api-access-7vvdz" (OuterVolumeSpecName: "kube-api-access-7vvdz") pod "ebb5441f-839a-408b-bd51-3460006c3d83" (UID: "ebb5441f-839a-408b-bd51-3460006c3d83"). InnerVolumeSpecName "kube-api-access-7vvdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.063778 4959 generic.go:334] "Generic (PLEG): container finished" podID="cb8de785-a298-414c-998c-7bd0585966e1" containerID="0127c5ab88c1d7a1993c76d818a43a226a5ee407b5f5aaa3ff686cc1d404aa63" exitCode=0 Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.063820 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xwlzm" event={"ID":"cb8de785-a298-414c-998c-7bd0585966e1","Type":"ContainerDied","Data":"0127c5ab88c1d7a1993c76d818a43a226a5ee407b5f5aaa3ff686cc1d404aa63"} Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.063848 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xwlzm" event={"ID":"cb8de785-a298-414c-998c-7bd0585966e1","Type":"ContainerDied","Data":"b5c6f3bda831fa6d2461f11a9f68d20440ee425b5e270cb2bd2bb004178e7670"} Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.064238 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xwlzm" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.080572 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c94wx" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.096049 4959 scope.go:117] "RemoveContainer" containerID="b1ac98632af547caf91e4d584980a4a3bc41aa0808cab60cffc1cb44355c9fbb" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.116369 4959 scope.go:117] "RemoveContainer" containerID="2daa6f510281017ba2c87047c2b8f0b82762823189e3d76885a628912d0e67a0" Oct 07 13:49:45 crc kubenswrapper[4959]: E1007 13:49:45.118457 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2daa6f510281017ba2c87047c2b8f0b82762823189e3d76885a628912d0e67a0\": container with ID starting with 2daa6f510281017ba2c87047c2b8f0b82762823189e3d76885a628912d0e67a0 not found: ID does not exist" containerID="2daa6f510281017ba2c87047c2b8f0b82762823189e3d76885a628912d0e67a0" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.118515 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2daa6f510281017ba2c87047c2b8f0b82762823189e3d76885a628912d0e67a0"} err="failed to get container status \"2daa6f510281017ba2c87047c2b8f0b82762823189e3d76885a628912d0e67a0\": rpc error: code = NotFound desc = could not find container \"2daa6f510281017ba2c87047c2b8f0b82762823189e3d76885a628912d0e67a0\": container with ID starting with 2daa6f510281017ba2c87047c2b8f0b82762823189e3d76885a628912d0e67a0 not found: ID does not exist" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.118553 4959 scope.go:117] "RemoveContainer" containerID="00a246533cb95ee3510722163a228aba26a544d17b441a6cb7e496d9ea99c053" Oct 07 13:49:45 crc kubenswrapper[4959]: E1007 13:49:45.118878 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00a246533cb95ee3510722163a228aba26a544d17b441a6cb7e496d9ea99c053\": container with ID starting with 00a246533cb95ee3510722163a228aba26a544d17b441a6cb7e496d9ea99c053 not found: ID does not exist" containerID="00a246533cb95ee3510722163a228aba26a544d17b441a6cb7e496d9ea99c053" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.118912 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00a246533cb95ee3510722163a228aba26a544d17b441a6cb7e496d9ea99c053"} err="failed to get container status \"00a246533cb95ee3510722163a228aba26a544d17b441a6cb7e496d9ea99c053\": rpc error: code = NotFound desc = could not find container \"00a246533cb95ee3510722163a228aba26a544d17b441a6cb7e496d9ea99c053\": container with ID starting with 00a246533cb95ee3510722163a228aba26a544d17b441a6cb7e496d9ea99c053 not found: ID does not exist" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.118954 4959 scope.go:117] "RemoveContainer" containerID="b1ac98632af547caf91e4d584980a4a3bc41aa0808cab60cffc1cb44355c9fbb" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.123299 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5r424"] Oct 07 13:49:45 crc kubenswrapper[4959]: E1007 13:49:45.124874 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1ac98632af547caf91e4d584980a4a3bc41aa0808cab60cffc1cb44355c9fbb\": container with ID starting with b1ac98632af547caf91e4d584980a4a3bc41aa0808cab60cffc1cb44355c9fbb not found: ID does not exist" containerID="b1ac98632af547caf91e4d584980a4a3bc41aa0808cab60cffc1cb44355c9fbb" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.124902 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1ac98632af547caf91e4d584980a4a3bc41aa0808cab60cffc1cb44355c9fbb"} err="failed to get container status \"b1ac98632af547caf91e4d584980a4a3bc41aa0808cab60cffc1cb44355c9fbb\": rpc error: code = NotFound desc = could not find container \"b1ac98632af547caf91e4d584980a4a3bc41aa0808cab60cffc1cb44355c9fbb\": container with ID starting with b1ac98632af547caf91e4d584980a4a3bc41aa0808cab60cffc1cb44355c9fbb not found: ID does not exist" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.124921 4959 scope.go:117] "RemoveContainer" containerID="b2384da6a2e7f871634616e179a6b68fbbec8011b37cc0931b7babc6c27c9546" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.125901 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5r424"] Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.136738 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-925d4\" (UniqueName: \"kubernetes.io/projected/44b4f9d4-243a-46a8-ab8a-4fb5ae11816b-kube-api-access-925d4\") pod \"44b4f9d4-243a-46a8-ab8a-4fb5ae11816b\" (UID: \"44b4f9d4-243a-46a8-ab8a-4fb5ae11816b\") " Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.136799 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44b4f9d4-243a-46a8-ab8a-4fb5ae11816b-utilities\") pod \"44b4f9d4-243a-46a8-ab8a-4fb5ae11816b\" (UID: \"44b4f9d4-243a-46a8-ab8a-4fb5ae11816b\") " Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.136929 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44b4f9d4-243a-46a8-ab8a-4fb5ae11816b-catalog-content\") pod \"44b4f9d4-243a-46a8-ab8a-4fb5ae11816b\" (UID: \"44b4f9d4-243a-46a8-ab8a-4fb5ae11816b\") " Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.137155 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6d6vt\" (UniqueName: \"kubernetes.io/projected/cb8de785-a298-414c-998c-7bd0585966e1-kube-api-access-6d6vt\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.137168 4959 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/cb8de785-a298-414c-998c-7bd0585966e1-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.137177 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zw8q2\" (UniqueName: \"kubernetes.io/projected/339283f2-dc57-4316-a334-6d79fbf2be9b-kube-api-access-zw8q2\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.137185 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebb5441f-839a-408b-bd51-3460006c3d83-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.137193 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/339283f2-dc57-4316-a334-6d79fbf2be9b-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.137202 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vvdz\" (UniqueName: \"kubernetes.io/projected/ebb5441f-839a-408b-bd51-3460006c3d83-kube-api-access-7vvdz\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.137210 4959 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cb8de785-a298-414c-998c-7bd0585966e1-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.140560 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44b4f9d4-243a-46a8-ab8a-4fb5ae11816b-utilities" (OuterVolumeSpecName: "utilities") pod "44b4f9d4-243a-46a8-ab8a-4fb5ae11816b" (UID: "44b4f9d4-243a-46a8-ab8a-4fb5ae11816b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.148305 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44b4f9d4-243a-46a8-ab8a-4fb5ae11816b-kube-api-access-925d4" (OuterVolumeSpecName: "kube-api-access-925d4") pod "44b4f9d4-243a-46a8-ab8a-4fb5ae11816b" (UID: "44b4f9d4-243a-46a8-ab8a-4fb5ae11816b"). InnerVolumeSpecName "kube-api-access-925d4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.151130 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebb5441f-839a-408b-bd51-3460006c3d83-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ebb5441f-839a-408b-bd51-3460006c3d83" (UID: "ebb5441f-839a-408b-bd51-3460006c3d83"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.153205 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xwlzm"] Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.153205 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44b4f9d4-243a-46a8-ab8a-4fb5ae11816b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "44b4f9d4-243a-46a8-ab8a-4fb5ae11816b" (UID: "44b4f9d4-243a-46a8-ab8a-4fb5ae11816b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.155463 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/339283f2-dc57-4316-a334-6d79fbf2be9b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "339283f2-dc57-4316-a334-6d79fbf2be9b" (UID: "339283f2-dc57-4316-a334-6d79fbf2be9b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.155504 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xwlzm"] Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.159107 4959 scope.go:117] "RemoveContainer" containerID="109edd1598c60b297dfb00054814dabc127bddcf64ba1f23423a35004fa1fc76" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.175049 4959 scope.go:117] "RemoveContainer" containerID="1f490ddca97ee825004567944d6c8291950ca819b496df51ca9bcb47e877e88d" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.187047 4959 scope.go:117] "RemoveContainer" containerID="b2384da6a2e7f871634616e179a6b68fbbec8011b37cc0931b7babc6c27c9546" Oct 07 13:49:45 crc kubenswrapper[4959]: E1007 13:49:45.187416 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2384da6a2e7f871634616e179a6b68fbbec8011b37cc0931b7babc6c27c9546\": container with ID starting with b2384da6a2e7f871634616e179a6b68fbbec8011b37cc0931b7babc6c27c9546 not found: ID does not exist" containerID="b2384da6a2e7f871634616e179a6b68fbbec8011b37cc0931b7babc6c27c9546" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.187466 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2384da6a2e7f871634616e179a6b68fbbec8011b37cc0931b7babc6c27c9546"} err="failed to get container status \"b2384da6a2e7f871634616e179a6b68fbbec8011b37cc0931b7babc6c27c9546\": rpc error: code = NotFound desc = could not find container \"b2384da6a2e7f871634616e179a6b68fbbec8011b37cc0931b7babc6c27c9546\": container with ID starting with b2384da6a2e7f871634616e179a6b68fbbec8011b37cc0931b7babc6c27c9546 not found: ID does not exist" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.187492 4959 scope.go:117] "RemoveContainer" containerID="109edd1598c60b297dfb00054814dabc127bddcf64ba1f23423a35004fa1fc76" Oct 07 13:49:45 crc kubenswrapper[4959]: E1007 13:49:45.187793 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"109edd1598c60b297dfb00054814dabc127bddcf64ba1f23423a35004fa1fc76\": container with ID starting with 109edd1598c60b297dfb00054814dabc127bddcf64ba1f23423a35004fa1fc76 not found: ID does not exist" containerID="109edd1598c60b297dfb00054814dabc127bddcf64ba1f23423a35004fa1fc76" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.187851 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"109edd1598c60b297dfb00054814dabc127bddcf64ba1f23423a35004fa1fc76"} err="failed to get container status \"109edd1598c60b297dfb00054814dabc127bddcf64ba1f23423a35004fa1fc76\": rpc error: code = NotFound desc = could not find container \"109edd1598c60b297dfb00054814dabc127bddcf64ba1f23423a35004fa1fc76\": container with ID starting with 109edd1598c60b297dfb00054814dabc127bddcf64ba1f23423a35004fa1fc76 not found: ID does not exist" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.187889 4959 scope.go:117] "RemoveContainer" containerID="1f490ddca97ee825004567944d6c8291950ca819b496df51ca9bcb47e877e88d" Oct 07 13:49:45 crc kubenswrapper[4959]: E1007 13:49:45.188274 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f490ddca97ee825004567944d6c8291950ca819b496df51ca9bcb47e877e88d\": container with ID starting with 1f490ddca97ee825004567944d6c8291950ca819b496df51ca9bcb47e877e88d not found: ID does not exist" containerID="1f490ddca97ee825004567944d6c8291950ca819b496df51ca9bcb47e877e88d" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.188299 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f490ddca97ee825004567944d6c8291950ca819b496df51ca9bcb47e877e88d"} err="failed to get container status \"1f490ddca97ee825004567944d6c8291950ca819b496df51ca9bcb47e877e88d\": rpc error: code = NotFound desc = could not find container \"1f490ddca97ee825004567944d6c8291950ca819b496df51ca9bcb47e877e88d\": container with ID starting with 1f490ddca97ee825004567944d6c8291950ca819b496df51ca9bcb47e877e88d not found: ID does not exist" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.188314 4959 scope.go:117] "RemoveContainer" containerID="44ca7c831bca81e37b93e090d7dd71055c8d6605796726e271b35d43b47d707a" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.200434 4959 scope.go:117] "RemoveContainer" containerID="0d0836308b8ddf7f378e0a545fe3d6f67656e86ed37c3ba57f5fa83974e94e92" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.216429 4959 scope.go:117] "RemoveContainer" containerID="cd437d6fec756cb435f8df059a24775e70036e577fc980e0cb1fb9e38fbbeea7" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.238449 4959 scope.go:117] "RemoveContainer" containerID="44ca7c831bca81e37b93e090d7dd71055c8d6605796726e271b35d43b47d707a" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.238652 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44b4f9d4-243a-46a8-ab8a-4fb5ae11816b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.238681 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/339283f2-dc57-4316-a334-6d79fbf2be9b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.238694 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-925d4\" (UniqueName: \"kubernetes.io/projected/44b4f9d4-243a-46a8-ab8a-4fb5ae11816b-kube-api-access-925d4\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.238708 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44b4f9d4-243a-46a8-ab8a-4fb5ae11816b-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.238721 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebb5441f-839a-408b-bd51-3460006c3d83-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 13:49:45 crc kubenswrapper[4959]: E1007 13:49:45.238994 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44ca7c831bca81e37b93e090d7dd71055c8d6605796726e271b35d43b47d707a\": container with ID starting with 44ca7c831bca81e37b93e090d7dd71055c8d6605796726e271b35d43b47d707a not found: ID does not exist" containerID="44ca7c831bca81e37b93e090d7dd71055c8d6605796726e271b35d43b47d707a" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.239168 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44ca7c831bca81e37b93e090d7dd71055c8d6605796726e271b35d43b47d707a"} err="failed to get container status \"44ca7c831bca81e37b93e090d7dd71055c8d6605796726e271b35d43b47d707a\": rpc error: code = NotFound desc = could not find container \"44ca7c831bca81e37b93e090d7dd71055c8d6605796726e271b35d43b47d707a\": container with ID starting with 44ca7c831bca81e37b93e090d7dd71055c8d6605796726e271b35d43b47d707a not found: ID does not exist" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.239218 4959 scope.go:117] "RemoveContainer" containerID="0d0836308b8ddf7f378e0a545fe3d6f67656e86ed37c3ba57f5fa83974e94e92" Oct 07 13:49:45 crc kubenswrapper[4959]: E1007 13:49:45.239537 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d0836308b8ddf7f378e0a545fe3d6f67656e86ed37c3ba57f5fa83974e94e92\": container with ID starting with 0d0836308b8ddf7f378e0a545fe3d6f67656e86ed37c3ba57f5fa83974e94e92 not found: ID does not exist" containerID="0d0836308b8ddf7f378e0a545fe3d6f67656e86ed37c3ba57f5fa83974e94e92" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.239566 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d0836308b8ddf7f378e0a545fe3d6f67656e86ed37c3ba57f5fa83974e94e92"} err="failed to get container status \"0d0836308b8ddf7f378e0a545fe3d6f67656e86ed37c3ba57f5fa83974e94e92\": rpc error: code = NotFound desc = could not find container \"0d0836308b8ddf7f378e0a545fe3d6f67656e86ed37c3ba57f5fa83974e94e92\": container with ID starting with 0d0836308b8ddf7f378e0a545fe3d6f67656e86ed37c3ba57f5fa83974e94e92 not found: ID does not exist" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.239584 4959 scope.go:117] "RemoveContainer" containerID="cd437d6fec756cb435f8df059a24775e70036e577fc980e0cb1fb9e38fbbeea7" Oct 07 13:49:45 crc kubenswrapper[4959]: E1007 13:49:45.240133 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd437d6fec756cb435f8df059a24775e70036e577fc980e0cb1fb9e38fbbeea7\": container with ID starting with cd437d6fec756cb435f8df059a24775e70036e577fc980e0cb1fb9e38fbbeea7 not found: ID does not exist" containerID="cd437d6fec756cb435f8df059a24775e70036e577fc980e0cb1fb9e38fbbeea7" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.240161 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd437d6fec756cb435f8df059a24775e70036e577fc980e0cb1fb9e38fbbeea7"} err="failed to get container status \"cd437d6fec756cb435f8df059a24775e70036e577fc980e0cb1fb9e38fbbeea7\": rpc error: code = NotFound desc = could not find container \"cd437d6fec756cb435f8df059a24775e70036e577fc980e0cb1fb9e38fbbeea7\": container with ID starting with cd437d6fec756cb435f8df059a24775e70036e577fc980e0cb1fb9e38fbbeea7 not found: ID does not exist" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.240178 4959 scope.go:117] "RemoveContainer" containerID="0127c5ab88c1d7a1993c76d818a43a226a5ee407b5f5aaa3ff686cc1d404aa63" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.252642 4959 scope.go:117] "RemoveContainer" containerID="0127c5ab88c1d7a1993c76d818a43a226a5ee407b5f5aaa3ff686cc1d404aa63" Oct 07 13:49:45 crc kubenswrapper[4959]: E1007 13:49:45.253136 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0127c5ab88c1d7a1993c76d818a43a226a5ee407b5f5aaa3ff686cc1d404aa63\": container with ID starting with 0127c5ab88c1d7a1993c76d818a43a226a5ee407b5f5aaa3ff686cc1d404aa63 not found: ID does not exist" containerID="0127c5ab88c1d7a1993c76d818a43a226a5ee407b5f5aaa3ff686cc1d404aa63" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.253182 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0127c5ab88c1d7a1993c76d818a43a226a5ee407b5f5aaa3ff686cc1d404aa63"} err="failed to get container status \"0127c5ab88c1d7a1993c76d818a43a226a5ee407b5f5aaa3ff686cc1d404aa63\": rpc error: code = NotFound desc = could not find container \"0127c5ab88c1d7a1993c76d818a43a226a5ee407b5f5aaa3ff686cc1d404aa63\": container with ID starting with 0127c5ab88c1d7a1993c76d818a43a226a5ee407b5f5aaa3ff686cc1d404aa63 not found: ID does not exist" Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.283315 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-599kq"] Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.360689 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t7mrz"] Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.366183 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-t7mrz"] Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.374509 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d6vp9"] Oct 07 13:49:45 crc kubenswrapper[4959]: I1007 13:49:45.377230 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-d6vp9"] Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.070707 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c94wx" event={"ID":"44b4f9d4-243a-46a8-ab8a-4fb5ae11816b","Type":"ContainerDied","Data":"dee5554320b9e597f40ea1b85d874156840674b24f8986273a37c1e05c47ffac"} Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.070767 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c94wx" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.071007 4959 scope.go:117] "RemoveContainer" containerID="2948cb9150d3ce5ef0e539f41a521dbe6e02828b34f4ea21d9bb12593c8e4f6f" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.075267 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-599kq" event={"ID":"922b72b0-51d8-4914-b740-5be4725ab083","Type":"ContainerStarted","Data":"b353144b4ed250b6372272fd1f20575c4fc76340b89965e51b27327749596f02"} Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.075298 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-599kq" event={"ID":"922b72b0-51d8-4914-b740-5be4725ab083","Type":"ContainerStarted","Data":"8f5434053d359fadc302191b4cdcb74e85448b7777de18e4c30a206c0e1a3797"} Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.075622 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-599kq" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.078997 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-599kq" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.091261 4959 scope.go:117] "RemoveContainer" containerID="cd25de8a142f807182b6ed2b264dfffe32efe9fba00f41edc369445587a549da" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.097455 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-599kq" podStartSLOduration=2.097435158 podStartE2EDuration="2.097435158s" podCreationTimestamp="2025-10-07 13:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:49:46.09343383 +0000 UTC m=+268.176838175" watchObservedRunningTime="2025-10-07 13:49:46.097435158 +0000 UTC m=+268.180839483" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.111477 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c94wx"] Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.119344 4959 scope.go:117] "RemoveContainer" containerID="08ec0d46e8b05d2c9e1c30ce7036a48dc857a3f3dbe44b2690227f9074441034" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.122522 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-c94wx"] Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.611117 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5tthc"] Oct 07 13:49:46 crc kubenswrapper[4959]: E1007 13:49:46.611569 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="339283f2-dc57-4316-a334-6d79fbf2be9b" containerName="registry-server" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.611580 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="339283f2-dc57-4316-a334-6d79fbf2be9b" containerName="registry-server" Oct 07 13:49:46 crc kubenswrapper[4959]: E1007 13:49:46.611590 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44b4f9d4-243a-46a8-ab8a-4fb5ae11816b" containerName="registry-server" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.611596 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="44b4f9d4-243a-46a8-ab8a-4fb5ae11816b" containerName="registry-server" Oct 07 13:49:46 crc kubenswrapper[4959]: E1007 13:49:46.611606 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac" containerName="extract-content" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.611612 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac" containerName="extract-content" Oct 07 13:49:46 crc kubenswrapper[4959]: E1007 13:49:46.611620 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb8de785-a298-414c-998c-7bd0585966e1" containerName="marketplace-operator" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.611626 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb8de785-a298-414c-998c-7bd0585966e1" containerName="marketplace-operator" Oct 07 13:49:46 crc kubenswrapper[4959]: E1007 13:49:46.611633 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="339283f2-dc57-4316-a334-6d79fbf2be9b" containerName="extract-content" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.611638 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="339283f2-dc57-4316-a334-6d79fbf2be9b" containerName="extract-content" Oct 07 13:49:46 crc kubenswrapper[4959]: E1007 13:49:46.611647 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebb5441f-839a-408b-bd51-3460006c3d83" containerName="extract-utilities" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.611653 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebb5441f-839a-408b-bd51-3460006c3d83" containerName="extract-utilities" Oct 07 13:49:46 crc kubenswrapper[4959]: E1007 13:49:46.611660 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44b4f9d4-243a-46a8-ab8a-4fb5ae11816b" containerName="extract-content" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.611666 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="44b4f9d4-243a-46a8-ab8a-4fb5ae11816b" containerName="extract-content" Oct 07 13:49:46 crc kubenswrapper[4959]: E1007 13:49:46.611675 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac" containerName="extract-utilities" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.611681 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac" containerName="extract-utilities" Oct 07 13:49:46 crc kubenswrapper[4959]: E1007 13:49:46.611691 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44b4f9d4-243a-46a8-ab8a-4fb5ae11816b" containerName="extract-utilities" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.611696 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="44b4f9d4-243a-46a8-ab8a-4fb5ae11816b" containerName="extract-utilities" Oct 07 13:49:46 crc kubenswrapper[4959]: E1007 13:49:46.611702 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebb5441f-839a-408b-bd51-3460006c3d83" containerName="extract-content" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.611709 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebb5441f-839a-408b-bd51-3460006c3d83" containerName="extract-content" Oct 07 13:49:46 crc kubenswrapper[4959]: E1007 13:49:46.611718 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebb5441f-839a-408b-bd51-3460006c3d83" containerName="registry-server" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.611725 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebb5441f-839a-408b-bd51-3460006c3d83" containerName="registry-server" Oct 07 13:49:46 crc kubenswrapper[4959]: E1007 13:49:46.611733 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac" containerName="registry-server" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.611739 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac" containerName="registry-server" Oct 07 13:49:46 crc kubenswrapper[4959]: E1007 13:49:46.611749 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="339283f2-dc57-4316-a334-6d79fbf2be9b" containerName="extract-utilities" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.611756 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="339283f2-dc57-4316-a334-6d79fbf2be9b" containerName="extract-utilities" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.611845 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb8de785-a298-414c-998c-7bd0585966e1" containerName="marketplace-operator" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.611854 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="339283f2-dc57-4316-a334-6d79fbf2be9b" containerName="registry-server" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.611861 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac" containerName="registry-server" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.611869 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="44b4f9d4-243a-46a8-ab8a-4fb5ae11816b" containerName="registry-server" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.611876 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebb5441f-839a-408b-bd51-3460006c3d83" containerName="registry-server" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.616005 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5tthc" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.617957 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.622925 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5tthc"] Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.654967 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c-utilities\") pod \"redhat-marketplace-5tthc\" (UID: \"6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c\") " pod="openshift-marketplace/redhat-marketplace-5tthc" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.655134 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c-catalog-content\") pod \"redhat-marketplace-5tthc\" (UID: \"6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c\") " pod="openshift-marketplace/redhat-marketplace-5tthc" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.655203 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-676f9\" (UniqueName: \"kubernetes.io/projected/6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c-kube-api-access-676f9\") pod \"redhat-marketplace-5tthc\" (UID: \"6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c\") " pod="openshift-marketplace/redhat-marketplace-5tthc" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.661406 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="339283f2-dc57-4316-a334-6d79fbf2be9b" path="/var/lib/kubelet/pods/339283f2-dc57-4316-a334-6d79fbf2be9b/volumes" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.662160 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44b4f9d4-243a-46a8-ab8a-4fb5ae11816b" path="/var/lib/kubelet/pods/44b4f9d4-243a-46a8-ab8a-4fb5ae11816b/volumes" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.662711 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac" path="/var/lib/kubelet/pods/4b8a86d1-63b1-43cd-9a9d-798c40a5a1ac/volumes" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.663754 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb8de785-a298-414c-998c-7bd0585966e1" path="/var/lib/kubelet/pods/cb8de785-a298-414c-998c-7bd0585966e1/volumes" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.664214 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebb5441f-839a-408b-bd51-3460006c3d83" path="/var/lib/kubelet/pods/ebb5441f-839a-408b-bd51-3460006c3d83/volumes" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.756399 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c-utilities\") pod \"redhat-marketplace-5tthc\" (UID: \"6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c\") " pod="openshift-marketplace/redhat-marketplace-5tthc" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.756455 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c-catalog-content\") pod \"redhat-marketplace-5tthc\" (UID: \"6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c\") " pod="openshift-marketplace/redhat-marketplace-5tthc" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.756483 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-676f9\" (UniqueName: \"kubernetes.io/projected/6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c-kube-api-access-676f9\") pod \"redhat-marketplace-5tthc\" (UID: \"6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c\") " pod="openshift-marketplace/redhat-marketplace-5tthc" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.756997 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c-catalog-content\") pod \"redhat-marketplace-5tthc\" (UID: \"6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c\") " pod="openshift-marketplace/redhat-marketplace-5tthc" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.756999 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c-utilities\") pod \"redhat-marketplace-5tthc\" (UID: \"6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c\") " pod="openshift-marketplace/redhat-marketplace-5tthc" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.775970 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-676f9\" (UniqueName: \"kubernetes.io/projected/6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c-kube-api-access-676f9\") pod \"redhat-marketplace-5tthc\" (UID: \"6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c\") " pod="openshift-marketplace/redhat-marketplace-5tthc" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.813814 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-f6t5x"] Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.815473 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f6t5x" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.818339 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.823185 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f6t5x"] Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.857551 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sln78\" (UniqueName: \"kubernetes.io/projected/f1a1cb02-d50a-441e-8ec6-0d2a8a12b255-kube-api-access-sln78\") pod \"certified-operators-f6t5x\" (UID: \"f1a1cb02-d50a-441e-8ec6-0d2a8a12b255\") " pod="openshift-marketplace/certified-operators-f6t5x" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.857638 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1a1cb02-d50a-441e-8ec6-0d2a8a12b255-catalog-content\") pod \"certified-operators-f6t5x\" (UID: \"f1a1cb02-d50a-441e-8ec6-0d2a8a12b255\") " pod="openshift-marketplace/certified-operators-f6t5x" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.857667 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1a1cb02-d50a-441e-8ec6-0d2a8a12b255-utilities\") pod \"certified-operators-f6t5x\" (UID: \"f1a1cb02-d50a-441e-8ec6-0d2a8a12b255\") " pod="openshift-marketplace/certified-operators-f6t5x" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.937946 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5tthc" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.958742 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sln78\" (UniqueName: \"kubernetes.io/projected/f1a1cb02-d50a-441e-8ec6-0d2a8a12b255-kube-api-access-sln78\") pod \"certified-operators-f6t5x\" (UID: \"f1a1cb02-d50a-441e-8ec6-0d2a8a12b255\") " pod="openshift-marketplace/certified-operators-f6t5x" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.958792 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1a1cb02-d50a-441e-8ec6-0d2a8a12b255-catalog-content\") pod \"certified-operators-f6t5x\" (UID: \"f1a1cb02-d50a-441e-8ec6-0d2a8a12b255\") " pod="openshift-marketplace/certified-operators-f6t5x" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.958856 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1a1cb02-d50a-441e-8ec6-0d2a8a12b255-utilities\") pod \"certified-operators-f6t5x\" (UID: \"f1a1cb02-d50a-441e-8ec6-0d2a8a12b255\") " pod="openshift-marketplace/certified-operators-f6t5x" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.959496 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1a1cb02-d50a-441e-8ec6-0d2a8a12b255-utilities\") pod \"certified-operators-f6t5x\" (UID: \"f1a1cb02-d50a-441e-8ec6-0d2a8a12b255\") " pod="openshift-marketplace/certified-operators-f6t5x" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.960486 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1a1cb02-d50a-441e-8ec6-0d2a8a12b255-catalog-content\") pod \"certified-operators-f6t5x\" (UID: \"f1a1cb02-d50a-441e-8ec6-0d2a8a12b255\") " pod="openshift-marketplace/certified-operators-f6t5x" Oct 07 13:49:46 crc kubenswrapper[4959]: I1007 13:49:46.979555 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sln78\" (UniqueName: \"kubernetes.io/projected/f1a1cb02-d50a-441e-8ec6-0d2a8a12b255-kube-api-access-sln78\") pod \"certified-operators-f6t5x\" (UID: \"f1a1cb02-d50a-441e-8ec6-0d2a8a12b255\") " pod="openshift-marketplace/certified-operators-f6t5x" Oct 07 13:49:47 crc kubenswrapper[4959]: I1007 13:49:47.142720 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f6t5x" Oct 07 13:49:47 crc kubenswrapper[4959]: I1007 13:49:47.355006 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5tthc"] Oct 07 13:49:47 crc kubenswrapper[4959]: W1007 13:49:47.358435 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6fda9f88_41a8_47d3_bdcc_382a0ee9aa2c.slice/crio-bffb292245cfff97d5e1def01ded378937ad018eea578edc289946bc61d63206 WatchSource:0}: Error finding container bffb292245cfff97d5e1def01ded378937ad018eea578edc289946bc61d63206: Status 404 returned error can't find the container with id bffb292245cfff97d5e1def01ded378937ad018eea578edc289946bc61d63206 Oct 07 13:49:47 crc kubenswrapper[4959]: I1007 13:49:47.360291 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f6t5x"] Oct 07 13:49:47 crc kubenswrapper[4959]: W1007 13:49:47.363777 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf1a1cb02_d50a_441e_8ec6_0d2a8a12b255.slice/crio-857caa629fa34aed0e8d8c626600ece6fdf5e35fe9520d0fa20aa8cd41d0e7df WatchSource:0}: Error finding container 857caa629fa34aed0e8d8c626600ece6fdf5e35fe9520d0fa20aa8cd41d0e7df: Status 404 returned error can't find the container with id 857caa629fa34aed0e8d8c626600ece6fdf5e35fe9520d0fa20aa8cd41d0e7df Oct 07 13:49:48 crc kubenswrapper[4959]: I1007 13:49:48.098454 4959 generic.go:334] "Generic (PLEG): container finished" podID="f1a1cb02-d50a-441e-8ec6-0d2a8a12b255" containerID="51a81b5769f162334215d9101493fa7a61f0f787fa285594d20aa0c32f9b4979" exitCode=0 Oct 07 13:49:48 crc kubenswrapper[4959]: I1007 13:49:48.098558 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f6t5x" event={"ID":"f1a1cb02-d50a-441e-8ec6-0d2a8a12b255","Type":"ContainerDied","Data":"51a81b5769f162334215d9101493fa7a61f0f787fa285594d20aa0c32f9b4979"} Oct 07 13:49:48 crc kubenswrapper[4959]: I1007 13:49:48.098884 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f6t5x" event={"ID":"f1a1cb02-d50a-441e-8ec6-0d2a8a12b255","Type":"ContainerStarted","Data":"857caa629fa34aed0e8d8c626600ece6fdf5e35fe9520d0fa20aa8cd41d0e7df"} Oct 07 13:49:48 crc kubenswrapper[4959]: I1007 13:49:48.106167 4959 generic.go:334] "Generic (PLEG): container finished" podID="6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c" containerID="4df80fceabcbd5d98731ad15849fd43b0a6a2f4395a9339d592a653968cbc039" exitCode=0 Oct 07 13:49:48 crc kubenswrapper[4959]: I1007 13:49:48.106283 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5tthc" event={"ID":"6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c","Type":"ContainerDied","Data":"4df80fceabcbd5d98731ad15849fd43b0a6a2f4395a9339d592a653968cbc039"} Oct 07 13:49:48 crc kubenswrapper[4959]: I1007 13:49:48.106317 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5tthc" event={"ID":"6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c","Type":"ContainerStarted","Data":"bffb292245cfff97d5e1def01ded378937ad018eea578edc289946bc61d63206"} Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.016620 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jhqgr"] Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.017780 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jhqgr" Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.021314 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.027702 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jhqgr"] Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.083059 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5g9n\" (UniqueName: \"kubernetes.io/projected/2f98e86e-cc03-4344-82d6-897514556718-kube-api-access-w5g9n\") pod \"redhat-operators-jhqgr\" (UID: \"2f98e86e-cc03-4344-82d6-897514556718\") " pod="openshift-marketplace/redhat-operators-jhqgr" Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.083124 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f98e86e-cc03-4344-82d6-897514556718-catalog-content\") pod \"redhat-operators-jhqgr\" (UID: \"2f98e86e-cc03-4344-82d6-897514556718\") " pod="openshift-marketplace/redhat-operators-jhqgr" Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.083152 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f98e86e-cc03-4344-82d6-897514556718-utilities\") pod \"redhat-operators-jhqgr\" (UID: \"2f98e86e-cc03-4344-82d6-897514556718\") " pod="openshift-marketplace/redhat-operators-jhqgr" Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.114193 4959 generic.go:334] "Generic (PLEG): container finished" podID="6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c" containerID="0dbca168f98b8bd135cd374ca60ea56bf608d9f6bdabf77b8cb7d2523472150d" exitCode=0 Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.114264 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5tthc" event={"ID":"6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c","Type":"ContainerDied","Data":"0dbca168f98b8bd135cd374ca60ea56bf608d9f6bdabf77b8cb7d2523472150d"} Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.117718 4959 generic.go:334] "Generic (PLEG): container finished" podID="f1a1cb02-d50a-441e-8ec6-0d2a8a12b255" containerID="3ee358ca133d56ccf7ea430882408f812f1b3e4154fe4396ba81cd80bdf651e5" exitCode=0 Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.117738 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f6t5x" event={"ID":"f1a1cb02-d50a-441e-8ec6-0d2a8a12b255","Type":"ContainerDied","Data":"3ee358ca133d56ccf7ea430882408f812f1b3e4154fe4396ba81cd80bdf651e5"} Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.184566 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f98e86e-cc03-4344-82d6-897514556718-utilities\") pod \"redhat-operators-jhqgr\" (UID: \"2f98e86e-cc03-4344-82d6-897514556718\") " pod="openshift-marketplace/redhat-operators-jhqgr" Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.184702 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5g9n\" (UniqueName: \"kubernetes.io/projected/2f98e86e-cc03-4344-82d6-897514556718-kube-api-access-w5g9n\") pod \"redhat-operators-jhqgr\" (UID: \"2f98e86e-cc03-4344-82d6-897514556718\") " pod="openshift-marketplace/redhat-operators-jhqgr" Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.184742 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f98e86e-cc03-4344-82d6-897514556718-catalog-content\") pod \"redhat-operators-jhqgr\" (UID: \"2f98e86e-cc03-4344-82d6-897514556718\") " pod="openshift-marketplace/redhat-operators-jhqgr" Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.185985 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f98e86e-cc03-4344-82d6-897514556718-utilities\") pod \"redhat-operators-jhqgr\" (UID: \"2f98e86e-cc03-4344-82d6-897514556718\") " pod="openshift-marketplace/redhat-operators-jhqgr" Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.186090 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f98e86e-cc03-4344-82d6-897514556718-catalog-content\") pod \"redhat-operators-jhqgr\" (UID: \"2f98e86e-cc03-4344-82d6-897514556718\") " pod="openshift-marketplace/redhat-operators-jhqgr" Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.225301 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sn6c8"] Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.227304 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sn6c8" Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.229448 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sn6c8"] Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.229752 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.229999 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5g9n\" (UniqueName: \"kubernetes.io/projected/2f98e86e-cc03-4344-82d6-897514556718-kube-api-access-w5g9n\") pod \"redhat-operators-jhqgr\" (UID: \"2f98e86e-cc03-4344-82d6-897514556718\") " pod="openshift-marketplace/redhat-operators-jhqgr" Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.285860 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26cb3091-1301-401a-9592-92ba50054762-utilities\") pod \"community-operators-sn6c8\" (UID: \"26cb3091-1301-401a-9592-92ba50054762\") " pod="openshift-marketplace/community-operators-sn6c8" Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.285956 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26cb3091-1301-401a-9592-92ba50054762-catalog-content\") pod \"community-operators-sn6c8\" (UID: \"26cb3091-1301-401a-9592-92ba50054762\") " pod="openshift-marketplace/community-operators-sn6c8" Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.285997 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sx5r\" (UniqueName: \"kubernetes.io/projected/26cb3091-1301-401a-9592-92ba50054762-kube-api-access-8sx5r\") pod \"community-operators-sn6c8\" (UID: \"26cb3091-1301-401a-9592-92ba50054762\") " pod="openshift-marketplace/community-operators-sn6c8" Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.333744 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jhqgr" Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.387402 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26cb3091-1301-401a-9592-92ba50054762-utilities\") pod \"community-operators-sn6c8\" (UID: \"26cb3091-1301-401a-9592-92ba50054762\") " pod="openshift-marketplace/community-operators-sn6c8" Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.387494 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26cb3091-1301-401a-9592-92ba50054762-catalog-content\") pod \"community-operators-sn6c8\" (UID: \"26cb3091-1301-401a-9592-92ba50054762\") " pod="openshift-marketplace/community-operators-sn6c8" Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.387537 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8sx5r\" (UniqueName: \"kubernetes.io/projected/26cb3091-1301-401a-9592-92ba50054762-kube-api-access-8sx5r\") pod \"community-operators-sn6c8\" (UID: \"26cb3091-1301-401a-9592-92ba50054762\") " pod="openshift-marketplace/community-operators-sn6c8" Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.387915 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26cb3091-1301-401a-9592-92ba50054762-utilities\") pod \"community-operators-sn6c8\" (UID: \"26cb3091-1301-401a-9592-92ba50054762\") " pod="openshift-marketplace/community-operators-sn6c8" Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.387996 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26cb3091-1301-401a-9592-92ba50054762-catalog-content\") pod \"community-operators-sn6c8\" (UID: \"26cb3091-1301-401a-9592-92ba50054762\") " pod="openshift-marketplace/community-operators-sn6c8" Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.409569 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8sx5r\" (UniqueName: \"kubernetes.io/projected/26cb3091-1301-401a-9592-92ba50054762-kube-api-access-8sx5r\") pod \"community-operators-sn6c8\" (UID: \"26cb3091-1301-401a-9592-92ba50054762\") " pod="openshift-marketplace/community-operators-sn6c8" Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.563953 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sn6c8" Oct 07 13:49:49 crc kubenswrapper[4959]: I1007 13:49:49.591826 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jhqgr"] Oct 07 13:49:50 crc kubenswrapper[4959]: I1007 13:49:50.008816 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sn6c8"] Oct 07 13:49:50 crc kubenswrapper[4959]: W1007 13:49:50.012135 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod26cb3091_1301_401a_9592_92ba50054762.slice/crio-f2aba2b41ba643beaa2e6e75f52093b364386da729fb1e33e48078e309d85e58 WatchSource:0}: Error finding container f2aba2b41ba643beaa2e6e75f52093b364386da729fb1e33e48078e309d85e58: Status 404 returned error can't find the container with id f2aba2b41ba643beaa2e6e75f52093b364386da729fb1e33e48078e309d85e58 Oct 07 13:49:50 crc kubenswrapper[4959]: I1007 13:49:50.124794 4959 generic.go:334] "Generic (PLEG): container finished" podID="2f98e86e-cc03-4344-82d6-897514556718" containerID="94d62cb480fba8ae245fb9eac6e93f7e178a06eb156c055d7ba5a6a1ef3bd06a" exitCode=0 Oct 07 13:49:50 crc kubenswrapper[4959]: I1007 13:49:50.125778 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhqgr" event={"ID":"2f98e86e-cc03-4344-82d6-897514556718","Type":"ContainerDied","Data":"94d62cb480fba8ae245fb9eac6e93f7e178a06eb156c055d7ba5a6a1ef3bd06a"} Oct 07 13:49:50 crc kubenswrapper[4959]: I1007 13:49:50.125847 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhqgr" event={"ID":"2f98e86e-cc03-4344-82d6-897514556718","Type":"ContainerStarted","Data":"42ad41d343ea16415faaf4b64918d0d04cb430cc56aa29853e571b4c00ac768d"} Oct 07 13:49:50 crc kubenswrapper[4959]: I1007 13:49:50.130243 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f6t5x" event={"ID":"f1a1cb02-d50a-441e-8ec6-0d2a8a12b255","Type":"ContainerStarted","Data":"b84a3561fec6eafb45bb30accb651de9437d81e3664414cd5b137f3a4a53d628"} Oct 07 13:49:50 crc kubenswrapper[4959]: I1007 13:49:50.139179 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5tthc" event={"ID":"6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c","Type":"ContainerStarted","Data":"84daa4bad363e1d9327d7a47a1926c84a885d4768bf9827b2ab782e960799403"} Oct 07 13:49:50 crc kubenswrapper[4959]: I1007 13:49:50.144234 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sn6c8" event={"ID":"26cb3091-1301-401a-9592-92ba50054762","Type":"ContainerStarted","Data":"f2aba2b41ba643beaa2e6e75f52093b364386da729fb1e33e48078e309d85e58"} Oct 07 13:49:50 crc kubenswrapper[4959]: I1007 13:49:50.166706 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5tthc" podStartSLOduration=2.691234489 podStartE2EDuration="4.166687854s" podCreationTimestamp="2025-10-07 13:49:46 +0000 UTC" firstStartedPulling="2025-10-07 13:49:48.107870622 +0000 UTC m=+270.191274947" lastFinishedPulling="2025-10-07 13:49:49.583323997 +0000 UTC m=+271.666728312" observedRunningTime="2025-10-07 13:49:50.1650496 +0000 UTC m=+272.248453935" watchObservedRunningTime="2025-10-07 13:49:50.166687854 +0000 UTC m=+272.250092179" Oct 07 13:49:50 crc kubenswrapper[4959]: I1007 13:49:50.180904 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-f6t5x" podStartSLOduration=2.5365263540000003 podStartE2EDuration="4.180885075s" podCreationTimestamp="2025-10-07 13:49:46 +0000 UTC" firstStartedPulling="2025-10-07 13:49:48.102333723 +0000 UTC m=+270.185738058" lastFinishedPulling="2025-10-07 13:49:49.746692454 +0000 UTC m=+271.830096779" observedRunningTime="2025-10-07 13:49:50.179679563 +0000 UTC m=+272.263083888" watchObservedRunningTime="2025-10-07 13:49:50.180885075 +0000 UTC m=+272.264289400" Oct 07 13:49:51 crc kubenswrapper[4959]: I1007 13:49:51.151077 4959 generic.go:334] "Generic (PLEG): container finished" podID="26cb3091-1301-401a-9592-92ba50054762" containerID="c44edaaed66251944b09f42656abf0e075b9ca6bae79625057091b5eb70c5469" exitCode=0 Oct 07 13:49:51 crc kubenswrapper[4959]: I1007 13:49:51.152054 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sn6c8" event={"ID":"26cb3091-1301-401a-9592-92ba50054762","Type":"ContainerDied","Data":"c44edaaed66251944b09f42656abf0e075b9ca6bae79625057091b5eb70c5469"} Oct 07 13:49:52 crc kubenswrapper[4959]: I1007 13:49:52.163859 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhqgr" event={"ID":"2f98e86e-cc03-4344-82d6-897514556718","Type":"ContainerStarted","Data":"79d577eb44ebdadd81340ebed939ea537a89e0860a636dfc426a659f16b3abad"} Oct 07 13:49:53 crc kubenswrapper[4959]: I1007 13:49:53.170032 4959 generic.go:334] "Generic (PLEG): container finished" podID="2f98e86e-cc03-4344-82d6-897514556718" containerID="79d577eb44ebdadd81340ebed939ea537a89e0860a636dfc426a659f16b3abad" exitCode=0 Oct 07 13:49:53 crc kubenswrapper[4959]: I1007 13:49:53.170136 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhqgr" event={"ID":"2f98e86e-cc03-4344-82d6-897514556718","Type":"ContainerDied","Data":"79d577eb44ebdadd81340ebed939ea537a89e0860a636dfc426a659f16b3abad"} Oct 07 13:49:54 crc kubenswrapper[4959]: I1007 13:49:54.181734 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhqgr" event={"ID":"2f98e86e-cc03-4344-82d6-897514556718","Type":"ContainerStarted","Data":"a24c3a555f2faa4755cf004da20929226b1ea34cc3fbaac4cd616ea3be782569"} Oct 07 13:49:54 crc kubenswrapper[4959]: I1007 13:49:54.183448 4959 generic.go:334] "Generic (PLEG): container finished" podID="26cb3091-1301-401a-9592-92ba50054762" containerID="3805f1381cbfeb9e10d2c865da516d4bdb3b94c73150654d5c1f24570606a487" exitCode=0 Oct 07 13:49:54 crc kubenswrapper[4959]: I1007 13:49:54.183485 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sn6c8" event={"ID":"26cb3091-1301-401a-9592-92ba50054762","Type":"ContainerDied","Data":"3805f1381cbfeb9e10d2c865da516d4bdb3b94c73150654d5c1f24570606a487"} Oct 07 13:49:54 crc kubenswrapper[4959]: I1007 13:49:54.221194 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jhqgr" podStartSLOduration=1.750702309 podStartE2EDuration="5.221175784s" podCreationTimestamp="2025-10-07 13:49:49 +0000 UTC" firstStartedPulling="2025-10-07 13:49:50.127838821 +0000 UTC m=+272.211243156" lastFinishedPulling="2025-10-07 13:49:53.598312306 +0000 UTC m=+275.681716631" observedRunningTime="2025-10-07 13:49:54.201770933 +0000 UTC m=+276.285175278" watchObservedRunningTime="2025-10-07 13:49:54.221175784 +0000 UTC m=+276.304580109" Oct 07 13:49:55 crc kubenswrapper[4959]: I1007 13:49:55.190756 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sn6c8" event={"ID":"26cb3091-1301-401a-9592-92ba50054762","Type":"ContainerStarted","Data":"8c45d6fd9b73033aa59078ceb12cca0d07400015f5de5869a59556c8d778f060"} Oct 07 13:49:55 crc kubenswrapper[4959]: I1007 13:49:55.259298 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sn6c8" podStartSLOduration=2.799356053 podStartE2EDuration="6.259271644s" podCreationTimestamp="2025-10-07 13:49:49 +0000 UTC" firstStartedPulling="2025-10-07 13:49:51.153682592 +0000 UTC m=+273.237086917" lastFinishedPulling="2025-10-07 13:49:54.613598173 +0000 UTC m=+276.697002508" observedRunningTime="2025-10-07 13:49:55.255727089 +0000 UTC m=+277.339131424" watchObservedRunningTime="2025-10-07 13:49:55.259271644 +0000 UTC m=+277.342675999" Oct 07 13:49:56 crc kubenswrapper[4959]: I1007 13:49:56.938039 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5tthc" Oct 07 13:49:56 crc kubenswrapper[4959]: I1007 13:49:56.938374 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5tthc" Oct 07 13:49:56 crc kubenswrapper[4959]: I1007 13:49:56.977172 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5tthc" Oct 07 13:49:57 crc kubenswrapper[4959]: I1007 13:49:57.143132 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-f6t5x" Oct 07 13:49:57 crc kubenswrapper[4959]: I1007 13:49:57.143203 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-f6t5x" Oct 07 13:49:57 crc kubenswrapper[4959]: I1007 13:49:57.178882 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-f6t5x" Oct 07 13:49:57 crc kubenswrapper[4959]: I1007 13:49:57.239577 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5tthc" Oct 07 13:49:57 crc kubenswrapper[4959]: I1007 13:49:57.241726 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-f6t5x" Oct 07 13:49:59 crc kubenswrapper[4959]: I1007 13:49:59.334908 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jhqgr" Oct 07 13:49:59 crc kubenswrapper[4959]: I1007 13:49:59.335395 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jhqgr" Oct 07 13:49:59 crc kubenswrapper[4959]: I1007 13:49:59.398417 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jhqgr" Oct 07 13:49:59 crc kubenswrapper[4959]: I1007 13:49:59.564587 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sn6c8" Oct 07 13:49:59 crc kubenswrapper[4959]: I1007 13:49:59.564688 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sn6c8" Oct 07 13:49:59 crc kubenswrapper[4959]: I1007 13:49:59.602182 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sn6c8" Oct 07 13:50:00 crc kubenswrapper[4959]: I1007 13:50:00.279433 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sn6c8" Oct 07 13:50:00 crc kubenswrapper[4959]: I1007 13:50:00.282701 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jhqgr" Oct 07 13:50:50 crc kubenswrapper[4959]: I1007 13:50:50.630215 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 13:50:50 crc kubenswrapper[4959]: I1007 13:50:50.630802 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 13:51:20 crc kubenswrapper[4959]: I1007 13:51:20.629796 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 13:51:20 crc kubenswrapper[4959]: I1007 13:51:20.630466 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 13:51:50 crc kubenswrapper[4959]: I1007 13:51:50.630693 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 13:51:50 crc kubenswrapper[4959]: I1007 13:51:50.631188 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 13:51:50 crc kubenswrapper[4959]: I1007 13:51:50.631234 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 13:51:50 crc kubenswrapper[4959]: I1007 13:51:50.631803 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ff34b703f977911ce787f57286ec64739600838a5ec5892759fc7837be39f505"} pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 13:51:50 crc kubenswrapper[4959]: I1007 13:51:50.631873 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" containerID="cri-o://ff34b703f977911ce787f57286ec64739600838a5ec5892759fc7837be39f505" gracePeriod=600 Oct 07 13:51:50 crc kubenswrapper[4959]: I1007 13:51:50.880767 4959 generic.go:334] "Generic (PLEG): container finished" podID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerID="ff34b703f977911ce787f57286ec64739600838a5ec5892759fc7837be39f505" exitCode=0 Oct 07 13:51:50 crc kubenswrapper[4959]: I1007 13:51:50.880831 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerDied","Data":"ff34b703f977911ce787f57286ec64739600838a5ec5892759fc7837be39f505"} Oct 07 13:51:50 crc kubenswrapper[4959]: I1007 13:51:50.881151 4959 scope.go:117] "RemoveContainer" containerID="bb06cd02a283bcbe09a5e1894593ead293bd6ffdf7ab14f8eaae83257019727f" Oct 07 13:51:51 crc kubenswrapper[4959]: I1007 13:51:51.888460 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerStarted","Data":"e4d224bb95cdc421b51fa1f72abe6b70cf3c87a3b06fde77476b65c5f022b080"} Oct 07 13:53:27 crc kubenswrapper[4959]: I1007 13:53:27.611206 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-z79mw"] Oct 07 13:53:27 crc kubenswrapper[4959]: I1007 13:53:27.612779 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" Oct 07 13:53:27 crc kubenswrapper[4959]: I1007 13:53:27.634430 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-z79mw"] Oct 07 13:53:27 crc kubenswrapper[4959]: I1007 13:53:27.764024 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/41c37434-456b-47d9-a36a-6ae56d3612a0-trusted-ca\") pod \"image-registry-66df7c8f76-z79mw\" (UID: \"41c37434-456b-47d9-a36a-6ae56d3612a0\") " pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" Oct 07 13:53:27 crc kubenswrapper[4959]: I1007 13:53:27.764139 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/41c37434-456b-47d9-a36a-6ae56d3612a0-ca-trust-extracted\") pod \"image-registry-66df7c8f76-z79mw\" (UID: \"41c37434-456b-47d9-a36a-6ae56d3612a0\") " pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" Oct 07 13:53:27 crc kubenswrapper[4959]: I1007 13:53:27.764309 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-892mq\" (UniqueName: \"kubernetes.io/projected/41c37434-456b-47d9-a36a-6ae56d3612a0-kube-api-access-892mq\") pod \"image-registry-66df7c8f76-z79mw\" (UID: \"41c37434-456b-47d9-a36a-6ae56d3612a0\") " pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" Oct 07 13:53:27 crc kubenswrapper[4959]: I1007 13:53:27.764383 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/41c37434-456b-47d9-a36a-6ae56d3612a0-bound-sa-token\") pod \"image-registry-66df7c8f76-z79mw\" (UID: \"41c37434-456b-47d9-a36a-6ae56d3612a0\") " pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" Oct 07 13:53:27 crc kubenswrapper[4959]: I1007 13:53:27.764418 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/41c37434-456b-47d9-a36a-6ae56d3612a0-registry-tls\") pod \"image-registry-66df7c8f76-z79mw\" (UID: \"41c37434-456b-47d9-a36a-6ae56d3612a0\") " pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" Oct 07 13:53:27 crc kubenswrapper[4959]: I1007 13:53:27.764484 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/41c37434-456b-47d9-a36a-6ae56d3612a0-registry-certificates\") pod \"image-registry-66df7c8f76-z79mw\" (UID: \"41c37434-456b-47d9-a36a-6ae56d3612a0\") " pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" Oct 07 13:53:27 crc kubenswrapper[4959]: I1007 13:53:27.764560 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/41c37434-456b-47d9-a36a-6ae56d3612a0-installation-pull-secrets\") pod \"image-registry-66df7c8f76-z79mw\" (UID: \"41c37434-456b-47d9-a36a-6ae56d3612a0\") " pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" Oct 07 13:53:27 crc kubenswrapper[4959]: I1007 13:53:27.764668 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-z79mw\" (UID: \"41c37434-456b-47d9-a36a-6ae56d3612a0\") " pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" Oct 07 13:53:27 crc kubenswrapper[4959]: I1007 13:53:27.789268 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-z79mw\" (UID: \"41c37434-456b-47d9-a36a-6ae56d3612a0\") " pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" Oct 07 13:53:27 crc kubenswrapper[4959]: I1007 13:53:27.866012 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/41c37434-456b-47d9-a36a-6ae56d3612a0-trusted-ca\") pod \"image-registry-66df7c8f76-z79mw\" (UID: \"41c37434-456b-47d9-a36a-6ae56d3612a0\") " pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" Oct 07 13:53:27 crc kubenswrapper[4959]: I1007 13:53:27.866089 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/41c37434-456b-47d9-a36a-6ae56d3612a0-ca-trust-extracted\") pod \"image-registry-66df7c8f76-z79mw\" (UID: \"41c37434-456b-47d9-a36a-6ae56d3612a0\") " pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" Oct 07 13:53:27 crc kubenswrapper[4959]: I1007 13:53:27.866153 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-892mq\" (UniqueName: \"kubernetes.io/projected/41c37434-456b-47d9-a36a-6ae56d3612a0-kube-api-access-892mq\") pod \"image-registry-66df7c8f76-z79mw\" (UID: \"41c37434-456b-47d9-a36a-6ae56d3612a0\") " pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" Oct 07 13:53:27 crc kubenswrapper[4959]: I1007 13:53:27.866178 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/41c37434-456b-47d9-a36a-6ae56d3612a0-bound-sa-token\") pod \"image-registry-66df7c8f76-z79mw\" (UID: \"41c37434-456b-47d9-a36a-6ae56d3612a0\") " pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" Oct 07 13:53:27 crc kubenswrapper[4959]: I1007 13:53:27.866233 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/41c37434-456b-47d9-a36a-6ae56d3612a0-registry-tls\") pod \"image-registry-66df7c8f76-z79mw\" (UID: \"41c37434-456b-47d9-a36a-6ae56d3612a0\") " pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" Oct 07 13:53:27 crc kubenswrapper[4959]: I1007 13:53:27.866264 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/41c37434-456b-47d9-a36a-6ae56d3612a0-registry-certificates\") pod \"image-registry-66df7c8f76-z79mw\" (UID: \"41c37434-456b-47d9-a36a-6ae56d3612a0\") " pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" Oct 07 13:53:27 crc kubenswrapper[4959]: I1007 13:53:27.866287 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/41c37434-456b-47d9-a36a-6ae56d3612a0-installation-pull-secrets\") pod \"image-registry-66df7c8f76-z79mw\" (UID: \"41c37434-456b-47d9-a36a-6ae56d3612a0\") " pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" Oct 07 13:53:27 crc kubenswrapper[4959]: I1007 13:53:27.867064 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/41c37434-456b-47d9-a36a-6ae56d3612a0-ca-trust-extracted\") pod \"image-registry-66df7c8f76-z79mw\" (UID: \"41c37434-456b-47d9-a36a-6ae56d3612a0\") " pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" Oct 07 13:53:27 crc kubenswrapper[4959]: I1007 13:53:27.867381 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/41c37434-456b-47d9-a36a-6ae56d3612a0-trusted-ca\") pod \"image-registry-66df7c8f76-z79mw\" (UID: \"41c37434-456b-47d9-a36a-6ae56d3612a0\") " pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" Oct 07 13:53:27 crc kubenswrapper[4959]: I1007 13:53:27.867986 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/41c37434-456b-47d9-a36a-6ae56d3612a0-registry-certificates\") pod \"image-registry-66df7c8f76-z79mw\" (UID: \"41c37434-456b-47d9-a36a-6ae56d3612a0\") " pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" Oct 07 13:53:27 crc kubenswrapper[4959]: I1007 13:53:27.871547 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/41c37434-456b-47d9-a36a-6ae56d3612a0-installation-pull-secrets\") pod \"image-registry-66df7c8f76-z79mw\" (UID: \"41c37434-456b-47d9-a36a-6ae56d3612a0\") " pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" Oct 07 13:53:27 crc kubenswrapper[4959]: I1007 13:53:27.874578 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/41c37434-456b-47d9-a36a-6ae56d3612a0-registry-tls\") pod \"image-registry-66df7c8f76-z79mw\" (UID: \"41c37434-456b-47d9-a36a-6ae56d3612a0\") " pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" Oct 07 13:53:27 crc kubenswrapper[4959]: I1007 13:53:27.880672 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/41c37434-456b-47d9-a36a-6ae56d3612a0-bound-sa-token\") pod \"image-registry-66df7c8f76-z79mw\" (UID: \"41c37434-456b-47d9-a36a-6ae56d3612a0\") " pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" Oct 07 13:53:27 crc kubenswrapper[4959]: I1007 13:53:27.894071 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-892mq\" (UniqueName: \"kubernetes.io/projected/41c37434-456b-47d9-a36a-6ae56d3612a0-kube-api-access-892mq\") pod \"image-registry-66df7c8f76-z79mw\" (UID: \"41c37434-456b-47d9-a36a-6ae56d3612a0\") " pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" Oct 07 13:53:27 crc kubenswrapper[4959]: I1007 13:53:27.932314 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" Oct 07 13:53:28 crc kubenswrapper[4959]: I1007 13:53:28.124534 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-z79mw"] Oct 07 13:53:28 crc kubenswrapper[4959]: I1007 13:53:28.511638 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" event={"ID":"41c37434-456b-47d9-a36a-6ae56d3612a0","Type":"ContainerStarted","Data":"b56aaf09e97a0454936d39ca88304a7cfebf2f3a8ef36e0029fba3524d7e3a56"} Oct 07 13:53:28 crc kubenswrapper[4959]: I1007 13:53:28.511689 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" event={"ID":"41c37434-456b-47d9-a36a-6ae56d3612a0","Type":"ContainerStarted","Data":"7a0d5b65978cdf7f32f65c6ead2be880667f53a114adf74dab93ee88e9f8700e"} Oct 07 13:53:28 crc kubenswrapper[4959]: I1007 13:53:28.511843 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" Oct 07 13:53:28 crc kubenswrapper[4959]: I1007 13:53:28.539404 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" podStartSLOduration=1.539385241 podStartE2EDuration="1.539385241s" podCreationTimestamp="2025-10-07 13:53:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:53:28.535771031 +0000 UTC m=+490.619175366" watchObservedRunningTime="2025-10-07 13:53:28.539385241 +0000 UTC m=+490.622789576" Oct 07 13:53:47 crc kubenswrapper[4959]: I1007 13:53:47.938812 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-z79mw" Oct 07 13:53:47 crc kubenswrapper[4959]: I1007 13:53:47.997826 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jhgsq"] Oct 07 13:53:50 crc kubenswrapper[4959]: I1007 13:53:50.630160 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 13:53:50 crc kubenswrapper[4959]: I1007 13:53:50.630544 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.051728 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" podUID="24ccfcc4-25a6-4264-86b1-6ba0097b2614" containerName="registry" containerID="cri-o://7e3b9068afcbcca93532dab4dc19a188ed50a6def356908e91711b5782b6ea4f" gracePeriod=30 Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.473816 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.583249 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/24ccfcc4-25a6-4264-86b1-6ba0097b2614-registry-certificates\") pod \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.583314 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/24ccfcc4-25a6-4264-86b1-6ba0097b2614-trusted-ca\") pod \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.583371 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/24ccfcc4-25a6-4264-86b1-6ba0097b2614-registry-tls\") pod \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.583560 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.583593 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/24ccfcc4-25a6-4264-86b1-6ba0097b2614-bound-sa-token\") pod \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.583635 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsm8h\" (UniqueName: \"kubernetes.io/projected/24ccfcc4-25a6-4264-86b1-6ba0097b2614-kube-api-access-xsm8h\") pod \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.583687 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/24ccfcc4-25a6-4264-86b1-6ba0097b2614-ca-trust-extracted\") pod \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.583722 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/24ccfcc4-25a6-4264-86b1-6ba0097b2614-installation-pull-secrets\") pod \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\" (UID: \"24ccfcc4-25a6-4264-86b1-6ba0097b2614\") " Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.584237 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24ccfcc4-25a6-4264-86b1-6ba0097b2614-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "24ccfcc4-25a6-4264-86b1-6ba0097b2614" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.584936 4959 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/24ccfcc4-25a6-4264-86b1-6ba0097b2614-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.585249 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24ccfcc4-25a6-4264-86b1-6ba0097b2614-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "24ccfcc4-25a6-4264-86b1-6ba0097b2614" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.589552 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24ccfcc4-25a6-4264-86b1-6ba0097b2614-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "24ccfcc4-25a6-4264-86b1-6ba0097b2614" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.590423 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24ccfcc4-25a6-4264-86b1-6ba0097b2614-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "24ccfcc4-25a6-4264-86b1-6ba0097b2614" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.590766 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24ccfcc4-25a6-4264-86b1-6ba0097b2614-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "24ccfcc4-25a6-4264-86b1-6ba0097b2614" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.591085 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24ccfcc4-25a6-4264-86b1-6ba0097b2614-kube-api-access-xsm8h" (OuterVolumeSpecName: "kube-api-access-xsm8h") pod "24ccfcc4-25a6-4264-86b1-6ba0097b2614" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614"). InnerVolumeSpecName "kube-api-access-xsm8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.594423 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "24ccfcc4-25a6-4264-86b1-6ba0097b2614" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.614368 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24ccfcc4-25a6-4264-86b1-6ba0097b2614-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "24ccfcc4-25a6-4264-86b1-6ba0097b2614" (UID: "24ccfcc4-25a6-4264-86b1-6ba0097b2614"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.685813 4959 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/24ccfcc4-25a6-4264-86b1-6ba0097b2614-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.686250 4959 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/24ccfcc4-25a6-4264-86b1-6ba0097b2614-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.686272 4959 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/24ccfcc4-25a6-4264-86b1-6ba0097b2614-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.686289 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsm8h\" (UniqueName: \"kubernetes.io/projected/24ccfcc4-25a6-4264-86b1-6ba0097b2614-kube-api-access-xsm8h\") on node \"crc\" DevicePath \"\"" Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.686306 4959 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/24ccfcc4-25a6-4264-86b1-6ba0097b2614-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.686323 4959 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/24ccfcc4-25a6-4264-86b1-6ba0097b2614-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.823765 4959 generic.go:334] "Generic (PLEG): container finished" podID="24ccfcc4-25a6-4264-86b1-6ba0097b2614" containerID="7e3b9068afcbcca93532dab4dc19a188ed50a6def356908e91711b5782b6ea4f" exitCode=0 Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.823818 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" event={"ID":"24ccfcc4-25a6-4264-86b1-6ba0097b2614","Type":"ContainerDied","Data":"7e3b9068afcbcca93532dab4dc19a188ed50a6def356908e91711b5782b6ea4f"} Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.823867 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" event={"ID":"24ccfcc4-25a6-4264-86b1-6ba0097b2614","Type":"ContainerDied","Data":"cdd43bc7abf0927692c795123bba03ab97784ecebbc50924418df50031e61079"} Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.823895 4959 scope.go:117] "RemoveContainer" containerID="7e3b9068afcbcca93532dab4dc19a188ed50a6def356908e91711b5782b6ea4f" Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.823900 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jhgsq" Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.849593 4959 scope.go:117] "RemoveContainer" containerID="7e3b9068afcbcca93532dab4dc19a188ed50a6def356908e91711b5782b6ea4f" Oct 07 13:54:13 crc kubenswrapper[4959]: E1007 13:54:13.851037 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e3b9068afcbcca93532dab4dc19a188ed50a6def356908e91711b5782b6ea4f\": container with ID starting with 7e3b9068afcbcca93532dab4dc19a188ed50a6def356908e91711b5782b6ea4f not found: ID does not exist" containerID="7e3b9068afcbcca93532dab4dc19a188ed50a6def356908e91711b5782b6ea4f" Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.851322 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e3b9068afcbcca93532dab4dc19a188ed50a6def356908e91711b5782b6ea4f"} err="failed to get container status \"7e3b9068afcbcca93532dab4dc19a188ed50a6def356908e91711b5782b6ea4f\": rpc error: code = NotFound desc = could not find container \"7e3b9068afcbcca93532dab4dc19a188ed50a6def356908e91711b5782b6ea4f\": container with ID starting with 7e3b9068afcbcca93532dab4dc19a188ed50a6def356908e91711b5782b6ea4f not found: ID does not exist" Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.872205 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jhgsq"] Oct 07 13:54:13 crc kubenswrapper[4959]: I1007 13:54:13.879117 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jhgsq"] Oct 07 13:54:14 crc kubenswrapper[4959]: I1007 13:54:14.668369 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24ccfcc4-25a6-4264-86b1-6ba0097b2614" path="/var/lib/kubelet/pods/24ccfcc4-25a6-4264-86b1-6ba0097b2614/volumes" Oct 07 13:54:20 crc kubenswrapper[4959]: I1007 13:54:20.630011 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 13:54:20 crc kubenswrapper[4959]: I1007 13:54:20.630534 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 13:54:50 crc kubenswrapper[4959]: I1007 13:54:50.630065 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 13:54:50 crc kubenswrapper[4959]: I1007 13:54:50.630656 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 13:54:50 crc kubenswrapper[4959]: I1007 13:54:50.630714 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 13:54:50 crc kubenswrapper[4959]: I1007 13:54:50.631461 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e4d224bb95cdc421b51fa1f72abe6b70cf3c87a3b06fde77476b65c5f022b080"} pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 13:54:50 crc kubenswrapper[4959]: I1007 13:54:50.631528 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" containerID="cri-o://e4d224bb95cdc421b51fa1f72abe6b70cf3c87a3b06fde77476b65c5f022b080" gracePeriod=600 Oct 07 13:54:51 crc kubenswrapper[4959]: I1007 13:54:51.055206 4959 generic.go:334] "Generic (PLEG): container finished" podID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerID="e4d224bb95cdc421b51fa1f72abe6b70cf3c87a3b06fde77476b65c5f022b080" exitCode=0 Oct 07 13:54:51 crc kubenswrapper[4959]: I1007 13:54:51.055564 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerDied","Data":"e4d224bb95cdc421b51fa1f72abe6b70cf3c87a3b06fde77476b65c5f022b080"} Oct 07 13:54:51 crc kubenswrapper[4959]: I1007 13:54:51.055601 4959 scope.go:117] "RemoveContainer" containerID="ff34b703f977911ce787f57286ec64739600838a5ec5892759fc7837be39f505" Oct 07 13:54:52 crc kubenswrapper[4959]: I1007 13:54:52.065433 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerStarted","Data":"59698f7bbe675b82b270fb68b1462a80d17b9a12f2121c8b8aa53d105e879f24"} Oct 07 13:55:25 crc kubenswrapper[4959]: I1007 13:55:25.632258 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-5d4td"] Oct 07 13:55:25 crc kubenswrapper[4959]: E1007 13:55:25.633017 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24ccfcc4-25a6-4264-86b1-6ba0097b2614" containerName="registry" Oct 07 13:55:25 crc kubenswrapper[4959]: I1007 13:55:25.633032 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="24ccfcc4-25a6-4264-86b1-6ba0097b2614" containerName="registry" Oct 07 13:55:25 crc kubenswrapper[4959]: I1007 13:55:25.633156 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="24ccfcc4-25a6-4264-86b1-6ba0097b2614" containerName="registry" Oct 07 13:55:25 crc kubenswrapper[4959]: I1007 13:55:25.633506 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-5d4td" Oct 07 13:55:25 crc kubenswrapper[4959]: I1007 13:55:25.636535 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 07 13:55:25 crc kubenswrapper[4959]: I1007 13:55:25.636955 4959 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-vgps5" Oct 07 13:55:25 crc kubenswrapper[4959]: I1007 13:55:25.637070 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 07 13:55:25 crc kubenswrapper[4959]: I1007 13:55:25.647027 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-h4qbt"] Oct 07 13:55:25 crc kubenswrapper[4959]: I1007 13:55:25.648028 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-h4qbt" Oct 07 13:55:25 crc kubenswrapper[4959]: I1007 13:55:25.650714 4959 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-h9x7k" Oct 07 13:55:25 crc kubenswrapper[4959]: I1007 13:55:25.653410 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-5d4td"] Oct 07 13:55:25 crc kubenswrapper[4959]: I1007 13:55:25.670372 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-pw7vl"] Oct 07 13:55:25 crc kubenswrapper[4959]: I1007 13:55:25.671521 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-pw7vl" Oct 07 13:55:25 crc kubenswrapper[4959]: I1007 13:55:25.673300 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2q5h\" (UniqueName: \"kubernetes.io/projected/0ab413ae-a8d2-41af-8abc-0f4169b3166e-kube-api-access-s2q5h\") pod \"cert-manager-webhook-5655c58dd6-pw7vl\" (UID: \"0ab413ae-a8d2-41af-8abc-0f4169b3166e\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-pw7vl" Oct 07 13:55:25 crc kubenswrapper[4959]: I1007 13:55:25.678577 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-h4qbt"] Oct 07 13:55:25 crc kubenswrapper[4959]: I1007 13:55:25.678728 4959 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-2pfzc" Oct 07 13:55:25 crc kubenswrapper[4959]: I1007 13:55:25.701200 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-pw7vl"] Oct 07 13:55:25 crc kubenswrapper[4959]: I1007 13:55:25.774393 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrz4b\" (UniqueName: \"kubernetes.io/projected/a0dc763a-75f9-41bf-bc04-f2da6c5bfc1e-kube-api-access-zrz4b\") pod \"cert-manager-cainjector-7f985d654d-5d4td\" (UID: \"a0dc763a-75f9-41bf-bc04-f2da6c5bfc1e\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-5d4td" Oct 07 13:55:25 crc kubenswrapper[4959]: I1007 13:55:25.774452 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6jn4\" (UniqueName: \"kubernetes.io/projected/b3b35c9d-c7b3-4baa-a282-464d7598355c-kube-api-access-m6jn4\") pod \"cert-manager-5b446d88c5-h4qbt\" (UID: \"b3b35c9d-c7b3-4baa-a282-464d7598355c\") " pod="cert-manager/cert-manager-5b446d88c5-h4qbt" Oct 07 13:55:25 crc kubenswrapper[4959]: I1007 13:55:25.774495 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2q5h\" (UniqueName: \"kubernetes.io/projected/0ab413ae-a8d2-41af-8abc-0f4169b3166e-kube-api-access-s2q5h\") pod \"cert-manager-webhook-5655c58dd6-pw7vl\" (UID: \"0ab413ae-a8d2-41af-8abc-0f4169b3166e\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-pw7vl" Oct 07 13:55:25 crc kubenswrapper[4959]: I1007 13:55:25.795558 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2q5h\" (UniqueName: \"kubernetes.io/projected/0ab413ae-a8d2-41af-8abc-0f4169b3166e-kube-api-access-s2q5h\") pod \"cert-manager-webhook-5655c58dd6-pw7vl\" (UID: \"0ab413ae-a8d2-41af-8abc-0f4169b3166e\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-pw7vl" Oct 07 13:55:25 crc kubenswrapper[4959]: I1007 13:55:25.875423 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrz4b\" (UniqueName: \"kubernetes.io/projected/a0dc763a-75f9-41bf-bc04-f2da6c5bfc1e-kube-api-access-zrz4b\") pod \"cert-manager-cainjector-7f985d654d-5d4td\" (UID: \"a0dc763a-75f9-41bf-bc04-f2da6c5bfc1e\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-5d4td" Oct 07 13:55:25 crc kubenswrapper[4959]: I1007 13:55:25.875755 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6jn4\" (UniqueName: \"kubernetes.io/projected/b3b35c9d-c7b3-4baa-a282-464d7598355c-kube-api-access-m6jn4\") pod \"cert-manager-5b446d88c5-h4qbt\" (UID: \"b3b35c9d-c7b3-4baa-a282-464d7598355c\") " pod="cert-manager/cert-manager-5b446d88c5-h4qbt" Oct 07 13:55:25 crc kubenswrapper[4959]: I1007 13:55:25.892720 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6jn4\" (UniqueName: \"kubernetes.io/projected/b3b35c9d-c7b3-4baa-a282-464d7598355c-kube-api-access-m6jn4\") pod \"cert-manager-5b446d88c5-h4qbt\" (UID: \"b3b35c9d-c7b3-4baa-a282-464d7598355c\") " pod="cert-manager/cert-manager-5b446d88c5-h4qbt" Oct 07 13:55:25 crc kubenswrapper[4959]: I1007 13:55:25.894058 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrz4b\" (UniqueName: \"kubernetes.io/projected/a0dc763a-75f9-41bf-bc04-f2da6c5bfc1e-kube-api-access-zrz4b\") pod \"cert-manager-cainjector-7f985d654d-5d4td\" (UID: \"a0dc763a-75f9-41bf-bc04-f2da6c5bfc1e\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-5d4td" Oct 07 13:55:25 crc kubenswrapper[4959]: I1007 13:55:25.959724 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-5d4td" Oct 07 13:55:25 crc kubenswrapper[4959]: I1007 13:55:25.966620 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-h4qbt" Oct 07 13:55:25 crc kubenswrapper[4959]: I1007 13:55:25.987562 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-pw7vl" Oct 07 13:55:26 crc kubenswrapper[4959]: I1007 13:55:26.232634 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-5d4td"] Oct 07 13:55:26 crc kubenswrapper[4959]: I1007 13:55:26.237305 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 13:55:26 crc kubenswrapper[4959]: I1007 13:55:26.257452 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-pw7vl"] Oct 07 13:55:26 crc kubenswrapper[4959]: W1007 13:55:26.258968 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ab413ae_a8d2_41af_8abc_0f4169b3166e.slice/crio-59d58780eaf65631dfb42f073f942ceef9b9e54df793b5dd127dc835272528b7 WatchSource:0}: Error finding container 59d58780eaf65631dfb42f073f942ceef9b9e54df793b5dd127dc835272528b7: Status 404 returned error can't find the container with id 59d58780eaf65631dfb42f073f942ceef9b9e54df793b5dd127dc835272528b7 Oct 07 13:55:26 crc kubenswrapper[4959]: I1007 13:55:26.296257 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-5d4td" event={"ID":"a0dc763a-75f9-41bf-bc04-f2da6c5bfc1e","Type":"ContainerStarted","Data":"ebd1bfde7bb6a63a8080cc2e8f444d7e7cacbc5a765568a9ee39274ebeefe3f0"} Oct 07 13:55:26 crc kubenswrapper[4959]: I1007 13:55:26.298850 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-pw7vl" event={"ID":"0ab413ae-a8d2-41af-8abc-0f4169b3166e","Type":"ContainerStarted","Data":"59d58780eaf65631dfb42f073f942ceef9b9e54df793b5dd127dc835272528b7"} Oct 07 13:55:26 crc kubenswrapper[4959]: I1007 13:55:26.409859 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-h4qbt"] Oct 07 13:55:27 crc kubenswrapper[4959]: I1007 13:55:27.312657 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-h4qbt" event={"ID":"b3b35c9d-c7b3-4baa-a282-464d7598355c","Type":"ContainerStarted","Data":"d65dc3ed597fb09476524879e2c404fa41b51e9c2edaf1683f97b618c3d0eaf9"} Oct 07 13:55:30 crc kubenswrapper[4959]: I1007 13:55:30.331807 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-5d4td" event={"ID":"a0dc763a-75f9-41bf-bc04-f2da6c5bfc1e","Type":"ContainerStarted","Data":"0f2b02707fd0122eec5d802ac5ccfd3cfb35848a707e93794a5f2a3090a4b4b3"} Oct 07 13:55:30 crc kubenswrapper[4959]: I1007 13:55:30.333398 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-pw7vl" event={"ID":"0ab413ae-a8d2-41af-8abc-0f4169b3166e","Type":"ContainerStarted","Data":"0bbb13f058e2b6c53814e0468fe1e15e146f2728d3342226d485d937ef138390"} Oct 07 13:55:30 crc kubenswrapper[4959]: I1007 13:55:30.333614 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-pw7vl" Oct 07 13:55:30 crc kubenswrapper[4959]: I1007 13:55:30.335026 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-h4qbt" event={"ID":"b3b35c9d-c7b3-4baa-a282-464d7598355c","Type":"ContainerStarted","Data":"4e8707801a45b8be085f6bd88aff970578db9c65ed3482f390b4ddc5b25b3bb2"} Oct 07 13:55:30 crc kubenswrapper[4959]: I1007 13:55:30.354600 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-5d4td" podStartSLOduration=2.073630263 podStartE2EDuration="5.354583811s" podCreationTimestamp="2025-10-07 13:55:25 +0000 UTC" firstStartedPulling="2025-10-07 13:55:26.237092619 +0000 UTC m=+608.320496944" lastFinishedPulling="2025-10-07 13:55:29.518046167 +0000 UTC m=+611.601450492" observedRunningTime="2025-10-07 13:55:30.35331005 +0000 UTC m=+612.436714425" watchObservedRunningTime="2025-10-07 13:55:30.354583811 +0000 UTC m=+612.437988136" Oct 07 13:55:30 crc kubenswrapper[4959]: I1007 13:55:30.384986 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-h4qbt" podStartSLOduration=2.280619154 podStartE2EDuration="5.384955441s" podCreationTimestamp="2025-10-07 13:55:25 +0000 UTC" firstStartedPulling="2025-10-07 13:55:26.41489156 +0000 UTC m=+608.498295885" lastFinishedPulling="2025-10-07 13:55:29.519227807 +0000 UTC m=+611.602632172" observedRunningTime="2025-10-07 13:55:30.377693167 +0000 UTC m=+612.461097532" watchObservedRunningTime="2025-10-07 13:55:30.384955441 +0000 UTC m=+612.468359816" Oct 07 13:55:30 crc kubenswrapper[4959]: I1007 13:55:30.400860 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-pw7vl" podStartSLOduration=2.162523294 podStartE2EDuration="5.400840963s" podCreationTimestamp="2025-10-07 13:55:25 +0000 UTC" firstStartedPulling="2025-10-07 13:55:26.260420608 +0000 UTC m=+608.343824933" lastFinishedPulling="2025-10-07 13:55:29.498738267 +0000 UTC m=+611.582142602" observedRunningTime="2025-10-07 13:55:30.395972189 +0000 UTC m=+612.479376524" watchObservedRunningTime="2025-10-07 13:55:30.400840963 +0000 UTC m=+612.484245288" Oct 07 13:55:35 crc kubenswrapper[4959]: I1007 13:55:35.991236 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-pw7vl" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.285924 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5sfv5"] Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.288485 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="ovn-controller" containerID="cri-o://1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f" gracePeriod=30 Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.288508 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="sbdb" containerID="cri-o://ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981" gracePeriod=30 Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.288627 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5" gracePeriod=30 Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.288677 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="kube-rbac-proxy-node" containerID="cri-o://d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3" gracePeriod=30 Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.288728 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="ovn-acl-logging" containerID="cri-o://f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a" gracePeriod=30 Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.288747 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="nbdb" containerID="cri-o://3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b" gracePeriod=30 Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.288794 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="northd" containerID="cri-o://bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c" gracePeriod=30 Oct 07 13:55:36 crc kubenswrapper[4959]: E1007 13:55:36.319806 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb79aa528_b2a2_4dff_b1d4_0e8dc918fc3b.slice/crio-d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3.scope\": RecentStats: unable to find data in memory cache]" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.339993 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="ovnkube-controller" containerID="cri-o://d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825" gracePeriod=30 Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.374506 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4mzl6_2003ec8f-74a0-47bc-8998-0326bfff6e7f/kube-multus/2.log" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.374860 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4mzl6_2003ec8f-74a0-47bc-8998-0326bfff6e7f/kube-multus/1.log" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.374893 4959 generic.go:334] "Generic (PLEG): container finished" podID="2003ec8f-74a0-47bc-8998-0326bfff6e7f" containerID="c030ef3529fc036b6e27c94b7c4c51d59bef679992157f58e1acf0858c581ed1" exitCode=2 Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.374917 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4mzl6" event={"ID":"2003ec8f-74a0-47bc-8998-0326bfff6e7f","Type":"ContainerDied","Data":"c030ef3529fc036b6e27c94b7c4c51d59bef679992157f58e1acf0858c581ed1"} Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.374948 4959 scope.go:117] "RemoveContainer" containerID="6898f5d8eb8c5a06853350d8189d1c393cef00392a0c6240b449dbb8617d9f2f" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.376222 4959 scope.go:117] "RemoveContainer" containerID="c030ef3529fc036b6e27c94b7c4c51d59bef679992157f58e1acf0858c581ed1" Oct 07 13:55:36 crc kubenswrapper[4959]: E1007 13:55:36.376510 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-4mzl6_openshift-multus(2003ec8f-74a0-47bc-8998-0326bfff6e7f)\"" pod="openshift-multus/multus-4mzl6" podUID="2003ec8f-74a0-47bc-8998-0326bfff6e7f" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.565401 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5sfv5_b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b/ovnkube-controller/3.log" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.567259 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5sfv5_b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b/ovn-acl-logging/0.log" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.567700 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5sfv5_b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b/ovn-controller/0.log" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.568162 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.625331 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4chhm"] Oct 07 13:55:36 crc kubenswrapper[4959]: E1007 13:55:36.625601 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="ovn-acl-logging" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.625618 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="ovn-acl-logging" Oct 07 13:55:36 crc kubenswrapper[4959]: E1007 13:55:36.625634 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="ovn-controller" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.625643 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="ovn-controller" Oct 07 13:55:36 crc kubenswrapper[4959]: E1007 13:55:36.625651 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="ovnkube-controller" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.625659 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="ovnkube-controller" Oct 07 13:55:36 crc kubenswrapper[4959]: E1007 13:55:36.625670 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="kubecfg-setup" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.625678 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="kubecfg-setup" Oct 07 13:55:36 crc kubenswrapper[4959]: E1007 13:55:36.625691 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="nbdb" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.625698 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="nbdb" Oct 07 13:55:36 crc kubenswrapper[4959]: E1007 13:55:36.625710 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="sbdb" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.625716 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="sbdb" Oct 07 13:55:36 crc kubenswrapper[4959]: E1007 13:55:36.625725 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="ovnkube-controller" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.625730 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="ovnkube-controller" Oct 07 13:55:36 crc kubenswrapper[4959]: E1007 13:55:36.625740 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="ovnkube-controller" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.625746 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="ovnkube-controller" Oct 07 13:55:36 crc kubenswrapper[4959]: E1007 13:55:36.625755 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="kube-rbac-proxy-node" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.625760 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="kube-rbac-proxy-node" Oct 07 13:55:36 crc kubenswrapper[4959]: E1007 13:55:36.625768 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="northd" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.625774 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="northd" Oct 07 13:55:36 crc kubenswrapper[4959]: E1007 13:55:36.625781 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="kube-rbac-proxy-ovn-metrics" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.625788 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="kube-rbac-proxy-ovn-metrics" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.625876 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="kube-rbac-proxy-ovn-metrics" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.625885 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="ovn-controller" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.625896 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="ovnkube-controller" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.625903 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="ovnkube-controller" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.625911 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="nbdb" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.625916 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="northd" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.625922 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="ovnkube-controller" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.625931 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="sbdb" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.625942 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="ovnkube-controller" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.625950 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="kube-rbac-proxy-node" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.625958 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="ovn-acl-logging" Oct 07 13:55:36 crc kubenswrapper[4959]: E1007 13:55:36.626047 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="ovnkube-controller" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.626053 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="ovnkube-controller" Oct 07 13:55:36 crc kubenswrapper[4959]: E1007 13:55:36.626065 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="ovnkube-controller" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.626071 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="ovnkube-controller" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.626178 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerName="ovnkube-controller" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.627752 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.661269 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-run-openvswitch\") pod \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.661346 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-969lv\" (UniqueName: \"kubernetes.io/projected/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-kube-api-access-969lv\") pod \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.661378 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-var-lib-openvswitch\") pod \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.661386 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" (UID: "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.661412 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-kubelet\") pod \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.661471 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-ovnkube-script-lib\") pod \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.661440 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" (UID: "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.661621 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" (UID: "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.661715 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-slash\") pod \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.661790 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-slash" (OuterVolumeSpecName: "host-slash") pod "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" (UID: "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.661886 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" (UID: "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.662808 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-etc-openvswitch\") pod \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.663005 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-systemd-units\") pod \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.662942 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" (UID: "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.663121 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" (UID: "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.663348 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-run-ovn\") pod \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.663419 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" (UID: "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.663466 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-node-log\") pod \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.663488 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-cni-bin\") pod \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.663552 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-node-log" (OuterVolumeSpecName: "node-log") pod "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" (UID: "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.663779 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-run-systemd\") pod \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.663838 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" (UID: "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.663895 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-run-ovn-kubernetes\") pod \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.664687 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" (UID: "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.664803 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-log-socket" (OuterVolumeSpecName: "log-socket") pod "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" (UID: "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.664833 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-log-socket\") pod \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.664872 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.664994 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-ovn-node-metrics-cert\") pod \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.665025 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-ovnkube-config\") pod \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.665043 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-cni-netd\") pod \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.665126 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-run-netns\") pod \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.665151 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-env-overrides\") pod \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\" (UID: \"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b\") " Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.665234 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" (UID: "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.665267 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" (UID: "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.665302 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" (UID: "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.665493 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" (UID: "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.665757 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" (UID: "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.665880 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-host-run-ovn-kubernetes\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.665959 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbn5n\" (UniqueName: \"kubernetes.io/projected/952d20e7-191a-47c0-b6b6-91cef18b8948-kube-api-access-qbn5n\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.665993 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/952d20e7-191a-47c0-b6b6-91cef18b8948-env-overrides\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.666068 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/952d20e7-191a-47c0-b6b6-91cef18b8948-ovnkube-script-lib\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.666144 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-host-run-netns\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.666177 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-host-slash\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.666246 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-log-socket\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.666269 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-var-lib-openvswitch\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.666325 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-node-log\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.666389 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-host-cni-bin\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.666417 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/952d20e7-191a-47c0-b6b6-91cef18b8948-ovn-node-metrics-cert\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.666494 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.666593 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/952d20e7-191a-47c0-b6b6-91cef18b8948-ovnkube-config\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.666663 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-etc-openvswitch\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.666756 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-run-systemd\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.666915 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-run-openvswitch\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.667044 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-host-cni-netd\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.667074 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-host-kubelet\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.667279 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-systemd-units\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.667433 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-run-ovn\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.667520 4959 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.667537 4959 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.667576 4959 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.667591 4959 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-slash\") on node \"crc\" DevicePath \"\"" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.667603 4959 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.667616 4959 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.667629 4959 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.667665 4959 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-node-log\") on node \"crc\" DevicePath \"\"" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.667677 4959 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.667691 4959 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.667704 4959 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-log-socket\") on node \"crc\" DevicePath \"\"" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.667718 4959 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.667733 4959 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.667746 4959 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.667759 4959 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.667777 4959 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.667790 4959 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.667838 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" (UID: "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.668264 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-kube-api-access-969lv" (OuterVolumeSpecName: "kube-api-access-969lv") pod "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" (UID: "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b"). InnerVolumeSpecName "kube-api-access-969lv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.677194 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" (UID: "b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769193 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-host-slash\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769243 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-log-socket\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769263 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-var-lib-openvswitch\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769284 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-node-log\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769302 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-host-cni-bin\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769316 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/952d20e7-191a-47c0-b6b6-91cef18b8948-ovn-node-metrics-cert\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769327 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-host-slash\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769361 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769335 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769397 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-node-log\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769408 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/952d20e7-191a-47c0-b6b6-91cef18b8948-ovnkube-config\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769421 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-host-cni-bin\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769414 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-var-lib-openvswitch\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769463 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-etc-openvswitch\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769431 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-log-socket\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769435 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-etc-openvswitch\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769507 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-run-systemd\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769532 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-run-openvswitch\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769586 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-host-cni-netd\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769610 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-host-kubelet\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769630 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-systemd-units\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769639 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-run-systemd\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769649 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-run-ovn\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769713 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-run-ovn\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769747 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-host-run-ovn-kubernetes\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769758 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-host-cni-netd\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769792 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-systemd-units\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769797 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-host-kubelet\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769715 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-host-run-ovn-kubernetes\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769832 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbn5n\" (UniqueName: \"kubernetes.io/projected/952d20e7-191a-47c0-b6b6-91cef18b8948-kube-api-access-qbn5n\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769838 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-run-openvswitch\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769856 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/952d20e7-191a-47c0-b6b6-91cef18b8948-env-overrides\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769898 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/952d20e7-191a-47c0-b6b6-91cef18b8948-ovnkube-script-lib\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769922 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-host-run-netns\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769971 4959 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.769991 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-969lv\" (UniqueName: \"kubernetes.io/projected/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-kube-api-access-969lv\") on node \"crc\" DevicePath \"\"" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.770005 4959 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.770046 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/952d20e7-191a-47c0-b6b6-91cef18b8948-host-run-netns\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.770704 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/952d20e7-191a-47c0-b6b6-91cef18b8948-env-overrides\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.770880 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/952d20e7-191a-47c0-b6b6-91cef18b8948-ovnkube-script-lib\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.771728 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/952d20e7-191a-47c0-b6b6-91cef18b8948-ovnkube-config\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.774144 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/952d20e7-191a-47c0-b6b6-91cef18b8948-ovn-node-metrics-cert\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.788751 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbn5n\" (UniqueName: \"kubernetes.io/projected/952d20e7-191a-47c0-b6b6-91cef18b8948-kube-api-access-qbn5n\") pod \"ovnkube-node-4chhm\" (UID: \"952d20e7-191a-47c0-b6b6-91cef18b8948\") " pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:36 crc kubenswrapper[4959]: I1007 13:55:36.941588 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.384229 4959 generic.go:334] "Generic (PLEG): container finished" podID="952d20e7-191a-47c0-b6b6-91cef18b8948" containerID="66dc5d3eb0162fbe8358e376498d6a425b99ea6ac8596528046cfc69da3b392c" exitCode=0 Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.384259 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" event={"ID":"952d20e7-191a-47c0-b6b6-91cef18b8948","Type":"ContainerDied","Data":"66dc5d3eb0162fbe8358e376498d6a425b99ea6ac8596528046cfc69da3b392c"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.384776 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" event={"ID":"952d20e7-191a-47c0-b6b6-91cef18b8948","Type":"ContainerStarted","Data":"3af75eb08568cdf8b61f1a35e4145fc14df7ae432bfa9e9b3936f2f1d19d79c9"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.390240 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5sfv5_b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b/ovnkube-controller/3.log" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.394677 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5sfv5_b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b/ovn-acl-logging/0.log" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.395827 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5sfv5_b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b/ovn-controller/0.log" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.396590 4959 generic.go:334] "Generic (PLEG): container finished" podID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerID="d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825" exitCode=0 Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.396641 4959 generic.go:334] "Generic (PLEG): container finished" podID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerID="ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981" exitCode=0 Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.396664 4959 generic.go:334] "Generic (PLEG): container finished" podID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerID="3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b" exitCode=0 Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.396681 4959 generic.go:334] "Generic (PLEG): container finished" podID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerID="bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c" exitCode=0 Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.396678 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" event={"ID":"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b","Type":"ContainerDied","Data":"d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.396763 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" event={"ID":"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b","Type":"ContainerDied","Data":"ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.396797 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" event={"ID":"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b","Type":"ContainerDied","Data":"3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.396695 4959 generic.go:334] "Generic (PLEG): container finished" podID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerID="7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5" exitCode=0 Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.396828 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" event={"ID":"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b","Type":"ContainerDied","Data":"bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.396862 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" event={"ID":"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b","Type":"ContainerDied","Data":"7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.396879 4959 scope.go:117] "RemoveContainer" containerID="d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.396890 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" event={"ID":"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b","Type":"ContainerDied","Data":"d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397055 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.396817 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397093 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397185 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397202 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397219 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.396832 4959 generic.go:334] "Generic (PLEG): container finished" podID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerID="d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3" exitCode=0 Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397275 4959 generic.go:334] "Generic (PLEG): container finished" podID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerID="f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a" exitCode=143 Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397305 4959 generic.go:334] "Generic (PLEG): container finished" podID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" containerID="1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f" exitCode=143 Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397233 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397390 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397413 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397428 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397484 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" event={"ID":"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b","Type":"ContainerDied","Data":"f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397517 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397529 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397540 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397551 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397562 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397573 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397584 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397594 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397605 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397616 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397631 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" event={"ID":"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b","Type":"ContainerDied","Data":"1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397648 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397662 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397677 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397688 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397703 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397714 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397758 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397769 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397780 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397791 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397806 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5sfv5" event={"ID":"b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b","Type":"ContainerDied","Data":"0fadf3af91429ed4e0681e8592293f1bbc052706c27bf35d235e0778d9bce43c"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397823 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397837 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397849 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397860 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397920 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397932 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397943 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.397954 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.398001 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.398016 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58"} Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.401195 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4mzl6_2003ec8f-74a0-47bc-8998-0326bfff6e7f/kube-multus/2.log" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.439947 4959 scope.go:117] "RemoveContainer" containerID="fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.496888 4959 scope.go:117] "RemoveContainer" containerID="ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.531855 4959 scope.go:117] "RemoveContainer" containerID="3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.544018 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5sfv5"] Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.558208 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5sfv5"] Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.559295 4959 scope.go:117] "RemoveContainer" containerID="bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.580719 4959 scope.go:117] "RemoveContainer" containerID="7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.601570 4959 scope.go:117] "RemoveContainer" containerID="d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.622171 4959 scope.go:117] "RemoveContainer" containerID="f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.640206 4959 scope.go:117] "RemoveContainer" containerID="1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.685286 4959 scope.go:117] "RemoveContainer" containerID="0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.714189 4959 scope.go:117] "RemoveContainer" containerID="d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825" Oct 07 13:55:37 crc kubenswrapper[4959]: E1007 13:55:37.714772 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825\": container with ID starting with d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825 not found: ID does not exist" containerID="d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.714807 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825"} err="failed to get container status \"d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825\": rpc error: code = NotFound desc = could not find container \"d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825\": container with ID starting with d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825 not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.714835 4959 scope.go:117] "RemoveContainer" containerID="fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69" Oct 07 13:55:37 crc kubenswrapper[4959]: E1007 13:55:37.715380 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69\": container with ID starting with fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69 not found: ID does not exist" containerID="fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.715399 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69"} err="failed to get container status \"fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69\": rpc error: code = NotFound desc = could not find container \"fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69\": container with ID starting with fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69 not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.715414 4959 scope.go:117] "RemoveContainer" containerID="ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981" Oct 07 13:55:37 crc kubenswrapper[4959]: E1007 13:55:37.715825 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\": container with ID starting with ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981 not found: ID does not exist" containerID="ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.715881 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981"} err="failed to get container status \"ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\": rpc error: code = NotFound desc = could not find container \"ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\": container with ID starting with ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981 not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.715931 4959 scope.go:117] "RemoveContainer" containerID="3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b" Oct 07 13:55:37 crc kubenswrapper[4959]: E1007 13:55:37.716265 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\": container with ID starting with 3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b not found: ID does not exist" containerID="3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.716285 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b"} err="failed to get container status \"3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\": rpc error: code = NotFound desc = could not find container \"3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\": container with ID starting with 3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.716299 4959 scope.go:117] "RemoveContainer" containerID="bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c" Oct 07 13:55:37 crc kubenswrapper[4959]: E1007 13:55:37.716572 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\": container with ID starting with bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c not found: ID does not exist" containerID="bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.716626 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c"} err="failed to get container status \"bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\": rpc error: code = NotFound desc = could not find container \"bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\": container with ID starting with bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.716662 4959 scope.go:117] "RemoveContainer" containerID="7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5" Oct 07 13:55:37 crc kubenswrapper[4959]: E1007 13:55:37.716956 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\": container with ID starting with 7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5 not found: ID does not exist" containerID="7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.716980 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5"} err="failed to get container status \"7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\": rpc error: code = NotFound desc = could not find container \"7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\": container with ID starting with 7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5 not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.716996 4959 scope.go:117] "RemoveContainer" containerID="d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3" Oct 07 13:55:37 crc kubenswrapper[4959]: E1007 13:55:37.717318 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\": container with ID starting with d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3 not found: ID does not exist" containerID="d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.717375 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3"} err="failed to get container status \"d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\": rpc error: code = NotFound desc = could not find container \"d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\": container with ID starting with d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3 not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.717410 4959 scope.go:117] "RemoveContainer" containerID="f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a" Oct 07 13:55:37 crc kubenswrapper[4959]: E1007 13:55:37.717667 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\": container with ID starting with f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a not found: ID does not exist" containerID="f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.717687 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a"} err="failed to get container status \"f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\": rpc error: code = NotFound desc = could not find container \"f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\": container with ID starting with f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.717702 4959 scope.go:117] "RemoveContainer" containerID="1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f" Oct 07 13:55:37 crc kubenswrapper[4959]: E1007 13:55:37.717936 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\": container with ID starting with 1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f not found: ID does not exist" containerID="1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.717953 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f"} err="failed to get container status \"1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\": rpc error: code = NotFound desc = could not find container \"1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\": container with ID starting with 1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.717967 4959 scope.go:117] "RemoveContainer" containerID="0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58" Oct 07 13:55:37 crc kubenswrapper[4959]: E1007 13:55:37.718274 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\": container with ID starting with 0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58 not found: ID does not exist" containerID="0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.718295 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58"} err="failed to get container status \"0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\": rpc error: code = NotFound desc = could not find container \"0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\": container with ID starting with 0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58 not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.718310 4959 scope.go:117] "RemoveContainer" containerID="d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.718567 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825"} err="failed to get container status \"d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825\": rpc error: code = NotFound desc = could not find container \"d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825\": container with ID starting with d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825 not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.718603 4959 scope.go:117] "RemoveContainer" containerID="fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.718874 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69"} err="failed to get container status \"fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69\": rpc error: code = NotFound desc = could not find container \"fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69\": container with ID starting with fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69 not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.718910 4959 scope.go:117] "RemoveContainer" containerID="ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.719181 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981"} err="failed to get container status \"ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\": rpc error: code = NotFound desc = could not find container \"ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\": container with ID starting with ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981 not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.719223 4959 scope.go:117] "RemoveContainer" containerID="3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.719558 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b"} err="failed to get container status \"3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\": rpc error: code = NotFound desc = could not find container \"3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\": container with ID starting with 3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.719596 4959 scope.go:117] "RemoveContainer" containerID="bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.719888 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c"} err="failed to get container status \"bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\": rpc error: code = NotFound desc = could not find container \"bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\": container with ID starting with bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.719920 4959 scope.go:117] "RemoveContainer" containerID="7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.720208 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5"} err="failed to get container status \"7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\": rpc error: code = NotFound desc = could not find container \"7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\": container with ID starting with 7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5 not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.720249 4959 scope.go:117] "RemoveContainer" containerID="d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.720509 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3"} err="failed to get container status \"d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\": rpc error: code = NotFound desc = could not find container \"d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\": container with ID starting with d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3 not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.720543 4959 scope.go:117] "RemoveContainer" containerID="f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.720793 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a"} err="failed to get container status \"f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\": rpc error: code = NotFound desc = could not find container \"f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\": container with ID starting with f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.720826 4959 scope.go:117] "RemoveContainer" containerID="1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.721080 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f"} err="failed to get container status \"1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\": rpc error: code = NotFound desc = could not find container \"1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\": container with ID starting with 1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.721140 4959 scope.go:117] "RemoveContainer" containerID="0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.721392 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58"} err="failed to get container status \"0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\": rpc error: code = NotFound desc = could not find container \"0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\": container with ID starting with 0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58 not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.721423 4959 scope.go:117] "RemoveContainer" containerID="d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.722536 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825"} err="failed to get container status \"d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825\": rpc error: code = NotFound desc = could not find container \"d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825\": container with ID starting with d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825 not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.722571 4959 scope.go:117] "RemoveContainer" containerID="fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.722897 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69"} err="failed to get container status \"fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69\": rpc error: code = NotFound desc = could not find container \"fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69\": container with ID starting with fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69 not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.722938 4959 scope.go:117] "RemoveContainer" containerID="ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.723217 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981"} err="failed to get container status \"ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\": rpc error: code = NotFound desc = could not find container \"ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\": container with ID starting with ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981 not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.723260 4959 scope.go:117] "RemoveContainer" containerID="3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.723827 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b"} err="failed to get container status \"3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\": rpc error: code = NotFound desc = could not find container \"3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\": container with ID starting with 3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.723871 4959 scope.go:117] "RemoveContainer" containerID="bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.724183 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c"} err="failed to get container status \"bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\": rpc error: code = NotFound desc = could not find container \"bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\": container with ID starting with bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.724225 4959 scope.go:117] "RemoveContainer" containerID="7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.724569 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5"} err="failed to get container status \"7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\": rpc error: code = NotFound desc = could not find container \"7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\": container with ID starting with 7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5 not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.724607 4959 scope.go:117] "RemoveContainer" containerID="d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.724865 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3"} err="failed to get container status \"d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\": rpc error: code = NotFound desc = could not find container \"d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\": container with ID starting with d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3 not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.724908 4959 scope.go:117] "RemoveContainer" containerID="f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.725423 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a"} err="failed to get container status \"f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\": rpc error: code = NotFound desc = could not find container \"f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\": container with ID starting with f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.725464 4959 scope.go:117] "RemoveContainer" containerID="1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.725742 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f"} err="failed to get container status \"1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\": rpc error: code = NotFound desc = could not find container \"1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\": container with ID starting with 1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.725783 4959 scope.go:117] "RemoveContainer" containerID="0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.726124 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58"} err="failed to get container status \"0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\": rpc error: code = NotFound desc = could not find container \"0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\": container with ID starting with 0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58 not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.726147 4959 scope.go:117] "RemoveContainer" containerID="d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.726427 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825"} err="failed to get container status \"d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825\": rpc error: code = NotFound desc = could not find container \"d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825\": container with ID starting with d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825 not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.726445 4959 scope.go:117] "RemoveContainer" containerID="fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.726817 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69"} err="failed to get container status \"fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69\": rpc error: code = NotFound desc = could not find container \"fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69\": container with ID starting with fcc39d7cda4afcd4baa1b14fd620a51afe3f0198dcedc5e55b2fe6c68c4eed69 not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.726862 4959 scope.go:117] "RemoveContainer" containerID="ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.727190 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981"} err="failed to get container status \"ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\": rpc error: code = NotFound desc = could not find container \"ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981\": container with ID starting with ea2d496f72381213907578f12e91d789be26aeb39c21d36110eaffd8cad70981 not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.727241 4959 scope.go:117] "RemoveContainer" containerID="3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.727581 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b"} err="failed to get container status \"3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\": rpc error: code = NotFound desc = could not find container \"3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b\": container with ID starting with 3053692ccc9e48930374d2ffa000de053a5617e4b320b709fda535da8180a04b not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.727623 4959 scope.go:117] "RemoveContainer" containerID="bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.727946 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c"} err="failed to get container status \"bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\": rpc error: code = NotFound desc = could not find container \"bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c\": container with ID starting with bdba3f7d72cd2ee0ac38926173d729a66b24ef78163b2dd95653e4e05a7afa3c not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.727996 4959 scope.go:117] "RemoveContainer" containerID="7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.728276 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5"} err="failed to get container status \"7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\": rpc error: code = NotFound desc = could not find container \"7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5\": container with ID starting with 7c92d38df4b8fde20da7cb1cccfcbc268b106e9eb3fedf33313fa3a0a48148c5 not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.728355 4959 scope.go:117] "RemoveContainer" containerID="d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.728636 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3"} err="failed to get container status \"d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\": rpc error: code = NotFound desc = could not find container \"d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3\": container with ID starting with d92d0a1a0295965ed22a39d4b87c011e72219adbb7b7a25dfc584889952c08b3 not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.728729 4959 scope.go:117] "RemoveContainer" containerID="f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.729047 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a"} err="failed to get container status \"f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\": rpc error: code = NotFound desc = could not find container \"f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a\": container with ID starting with f970245e513f7f1b1958f66f16e8eee1e4a6fb56321d707fe3c66dad91c77d3a not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.729089 4959 scope.go:117] "RemoveContainer" containerID="1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.729469 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f"} err="failed to get container status \"1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\": rpc error: code = NotFound desc = could not find container \"1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f\": container with ID starting with 1ca2b43e6de500abd6a7fc366f6d7ef109a5b574c771cbbf823cd6c4f723d54f not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.729520 4959 scope.go:117] "RemoveContainer" containerID="0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.729835 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58"} err="failed to get container status \"0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\": rpc error: code = NotFound desc = could not find container \"0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58\": container with ID starting with 0629244fed322f5430280d1534946d17c829ce60ae8bec799934188fc4970c58 not found: ID does not exist" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.729876 4959 scope.go:117] "RemoveContainer" containerID="d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825" Oct 07 13:55:37 crc kubenswrapper[4959]: I1007 13:55:37.730184 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825"} err="failed to get container status \"d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825\": rpc error: code = NotFound desc = could not find container \"d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825\": container with ID starting with d096903d95ad24b8975255f26663c0b79d547fc7c3702d62d0baa70e9906b825 not found: ID does not exist" Oct 07 13:55:38 crc kubenswrapper[4959]: I1007 13:55:38.411023 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" event={"ID":"952d20e7-191a-47c0-b6b6-91cef18b8948","Type":"ContainerStarted","Data":"0ce370e56b86e4ef33478a63716318958dc661b02646197d3ff12dcad918e1ec"} Oct 07 13:55:38 crc kubenswrapper[4959]: I1007 13:55:38.411125 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" event={"ID":"952d20e7-191a-47c0-b6b6-91cef18b8948","Type":"ContainerStarted","Data":"91d62072a4da69c2a4d49de3b21c24b157ec356d21b72cd4f3701a7bbdc1eaac"} Oct 07 13:55:38 crc kubenswrapper[4959]: I1007 13:55:38.411166 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" event={"ID":"952d20e7-191a-47c0-b6b6-91cef18b8948","Type":"ContainerStarted","Data":"153283d61de59a4bcb7eea68fbdd1d9a68f8277a59ded3f7b9ef598903234299"} Oct 07 13:55:38 crc kubenswrapper[4959]: I1007 13:55:38.411191 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" event={"ID":"952d20e7-191a-47c0-b6b6-91cef18b8948","Type":"ContainerStarted","Data":"b6b5058217777c3fb08fd1818365795d85ce2113cf759a1e42fdef2cec3f87bb"} Oct 07 13:55:38 crc kubenswrapper[4959]: I1007 13:55:38.411237 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" event={"ID":"952d20e7-191a-47c0-b6b6-91cef18b8948","Type":"ContainerStarted","Data":"75b512f02c135f82d76796504c41bce40c69d697a3aa646cf3c455b241979d0b"} Oct 07 13:55:38 crc kubenswrapper[4959]: I1007 13:55:38.411257 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" event={"ID":"952d20e7-191a-47c0-b6b6-91cef18b8948","Type":"ContainerStarted","Data":"64baf515651d033711787faaa9e1ed5e4218b3cdaa80fba32905f243dd95fbf1"} Oct 07 13:55:38 crc kubenswrapper[4959]: I1007 13:55:38.667234 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b" path="/var/lib/kubelet/pods/b79aa528-b2a2-4dff-b1d4-0e8dc918fc3b/volumes" Oct 07 13:55:41 crc kubenswrapper[4959]: I1007 13:55:41.449586 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" event={"ID":"952d20e7-191a-47c0-b6b6-91cef18b8948","Type":"ContainerStarted","Data":"81cabfeb5a7171e9931daa2976d7c1fabea0d659c20aff0d183e1d1640bafb2c"} Oct 07 13:55:43 crc kubenswrapper[4959]: I1007 13:55:43.468163 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" event={"ID":"952d20e7-191a-47c0-b6b6-91cef18b8948","Type":"ContainerStarted","Data":"ddb2e43cc52ca5e46760bbf650c1454b4b1cc48648115a18d1b9f725d9b00b98"} Oct 07 13:55:43 crc kubenswrapper[4959]: I1007 13:55:43.468638 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:43 crc kubenswrapper[4959]: I1007 13:55:43.468764 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:43 crc kubenswrapper[4959]: I1007 13:55:43.468791 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:43 crc kubenswrapper[4959]: I1007 13:55:43.493988 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:43 crc kubenswrapper[4959]: I1007 13:55:43.498216 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" podStartSLOduration=7.498200157 podStartE2EDuration="7.498200157s" podCreationTimestamp="2025-10-07 13:55:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:55:43.495975578 +0000 UTC m=+625.579379913" watchObservedRunningTime="2025-10-07 13:55:43.498200157 +0000 UTC m=+625.581604482" Oct 07 13:55:43 crc kubenswrapper[4959]: I1007 13:55:43.512274 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:55:50 crc kubenswrapper[4959]: I1007 13:55:50.653522 4959 scope.go:117] "RemoveContainer" containerID="c030ef3529fc036b6e27c94b7c4c51d59bef679992157f58e1acf0858c581ed1" Oct 07 13:55:50 crc kubenswrapper[4959]: E1007 13:55:50.654422 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-4mzl6_openshift-multus(2003ec8f-74a0-47bc-8998-0326bfff6e7f)\"" pod="openshift-multus/multus-4mzl6" podUID="2003ec8f-74a0-47bc-8998-0326bfff6e7f" Oct 07 13:56:01 crc kubenswrapper[4959]: I1007 13:56:01.653485 4959 scope.go:117] "RemoveContainer" containerID="c030ef3529fc036b6e27c94b7c4c51d59bef679992157f58e1acf0858c581ed1" Oct 07 13:56:02 crc kubenswrapper[4959]: I1007 13:56:02.609633 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4mzl6_2003ec8f-74a0-47bc-8998-0326bfff6e7f/kube-multus/2.log" Oct 07 13:56:02 crc kubenswrapper[4959]: I1007 13:56:02.609985 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4mzl6" event={"ID":"2003ec8f-74a0-47bc-8998-0326bfff6e7f","Type":"ContainerStarted","Data":"5825f7fc3c83f361c2834106fd435c7c52a9df16cba2e92a2a571c22d70e2a97"} Oct 07 13:56:06 crc kubenswrapper[4959]: I1007 13:56:06.979047 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4chhm" Oct 07 13:56:20 crc kubenswrapper[4959]: I1007 13:56:20.318268 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz"] Oct 07 13:56:20 crc kubenswrapper[4959]: I1007 13:56:20.320549 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz" Oct 07 13:56:20 crc kubenswrapper[4959]: I1007 13:56:20.324539 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 07 13:56:20 crc kubenswrapper[4959]: I1007 13:56:20.341653 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz"] Oct 07 13:56:20 crc kubenswrapper[4959]: I1007 13:56:20.396505 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfqfc\" (UniqueName: \"kubernetes.io/projected/bbe812c3-c65e-4a52-a87e-7138189e359a-kube-api-access-zfqfc\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz\" (UID: \"bbe812c3-c65e-4a52-a87e-7138189e359a\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz" Oct 07 13:56:20 crc kubenswrapper[4959]: I1007 13:56:20.396668 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bbe812c3-c65e-4a52-a87e-7138189e359a-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz\" (UID: \"bbe812c3-c65e-4a52-a87e-7138189e359a\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz" Oct 07 13:56:20 crc kubenswrapper[4959]: I1007 13:56:20.396700 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bbe812c3-c65e-4a52-a87e-7138189e359a-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz\" (UID: \"bbe812c3-c65e-4a52-a87e-7138189e359a\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz" Oct 07 13:56:20 crc kubenswrapper[4959]: I1007 13:56:20.498227 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bbe812c3-c65e-4a52-a87e-7138189e359a-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz\" (UID: \"bbe812c3-c65e-4a52-a87e-7138189e359a\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz" Oct 07 13:56:20 crc kubenswrapper[4959]: I1007 13:56:20.498287 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bbe812c3-c65e-4a52-a87e-7138189e359a-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz\" (UID: \"bbe812c3-c65e-4a52-a87e-7138189e359a\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz" Oct 07 13:56:20 crc kubenswrapper[4959]: I1007 13:56:20.498348 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfqfc\" (UniqueName: \"kubernetes.io/projected/bbe812c3-c65e-4a52-a87e-7138189e359a-kube-api-access-zfqfc\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz\" (UID: \"bbe812c3-c65e-4a52-a87e-7138189e359a\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz" Oct 07 13:56:20 crc kubenswrapper[4959]: I1007 13:56:20.498910 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bbe812c3-c65e-4a52-a87e-7138189e359a-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz\" (UID: \"bbe812c3-c65e-4a52-a87e-7138189e359a\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz" Oct 07 13:56:20 crc kubenswrapper[4959]: I1007 13:56:20.498972 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bbe812c3-c65e-4a52-a87e-7138189e359a-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz\" (UID: \"bbe812c3-c65e-4a52-a87e-7138189e359a\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz" Oct 07 13:56:20 crc kubenswrapper[4959]: I1007 13:56:20.523268 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfqfc\" (UniqueName: \"kubernetes.io/projected/bbe812c3-c65e-4a52-a87e-7138189e359a-kube-api-access-zfqfc\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz\" (UID: \"bbe812c3-c65e-4a52-a87e-7138189e359a\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz" Oct 07 13:56:20 crc kubenswrapper[4959]: I1007 13:56:20.652207 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz" Oct 07 13:56:21 crc kubenswrapper[4959]: I1007 13:56:21.073622 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz"] Oct 07 13:56:21 crc kubenswrapper[4959]: W1007 13:56:21.084257 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbbe812c3_c65e_4a52_a87e_7138189e359a.slice/crio-c2446f43de05328ded1c14725115a69a3a0156d8e49fb8295ab0ddd7f078d89b WatchSource:0}: Error finding container c2446f43de05328ded1c14725115a69a3a0156d8e49fb8295ab0ddd7f078d89b: Status 404 returned error can't find the container with id c2446f43de05328ded1c14725115a69a3a0156d8e49fb8295ab0ddd7f078d89b Oct 07 13:56:21 crc kubenswrapper[4959]: I1007 13:56:21.736247 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz" event={"ID":"bbe812c3-c65e-4a52-a87e-7138189e359a","Type":"ContainerStarted","Data":"9de63ef63e2cdf573313c352222d05cd15925ff5adf1cc4ddc902c37cb50d048"} Oct 07 13:56:21 crc kubenswrapper[4959]: I1007 13:56:21.736620 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz" event={"ID":"bbe812c3-c65e-4a52-a87e-7138189e359a","Type":"ContainerStarted","Data":"c2446f43de05328ded1c14725115a69a3a0156d8e49fb8295ab0ddd7f078d89b"} Oct 07 13:56:22 crc kubenswrapper[4959]: I1007 13:56:22.745884 4959 generic.go:334] "Generic (PLEG): container finished" podID="bbe812c3-c65e-4a52-a87e-7138189e359a" containerID="9de63ef63e2cdf573313c352222d05cd15925ff5adf1cc4ddc902c37cb50d048" exitCode=0 Oct 07 13:56:22 crc kubenswrapper[4959]: I1007 13:56:22.745957 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz" event={"ID":"bbe812c3-c65e-4a52-a87e-7138189e359a","Type":"ContainerDied","Data":"9de63ef63e2cdf573313c352222d05cd15925ff5adf1cc4ddc902c37cb50d048"} Oct 07 13:56:24 crc kubenswrapper[4959]: I1007 13:56:24.763880 4959 generic.go:334] "Generic (PLEG): container finished" podID="bbe812c3-c65e-4a52-a87e-7138189e359a" containerID="3ee7849168a7d3f5c35593d6e359d265bfd4bf7da883a2bfe41e5f3a550cf9d4" exitCode=0 Oct 07 13:56:24 crc kubenswrapper[4959]: I1007 13:56:24.763987 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz" event={"ID":"bbe812c3-c65e-4a52-a87e-7138189e359a","Type":"ContainerDied","Data":"3ee7849168a7d3f5c35593d6e359d265bfd4bf7da883a2bfe41e5f3a550cf9d4"} Oct 07 13:56:25 crc kubenswrapper[4959]: I1007 13:56:25.774778 4959 generic.go:334] "Generic (PLEG): container finished" podID="bbe812c3-c65e-4a52-a87e-7138189e359a" containerID="856cfd2a4eb7e7c2ce4aba2edb96cf28b372dc6f0cd6cac03709a5613aa84f56" exitCode=0 Oct 07 13:56:25 crc kubenswrapper[4959]: I1007 13:56:25.774844 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz" event={"ID":"bbe812c3-c65e-4a52-a87e-7138189e359a","Type":"ContainerDied","Data":"856cfd2a4eb7e7c2ce4aba2edb96cf28b372dc6f0cd6cac03709a5613aa84f56"} Oct 07 13:56:27 crc kubenswrapper[4959]: I1007 13:56:27.077010 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz" Oct 07 13:56:27 crc kubenswrapper[4959]: I1007 13:56:27.194433 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfqfc\" (UniqueName: \"kubernetes.io/projected/bbe812c3-c65e-4a52-a87e-7138189e359a-kube-api-access-zfqfc\") pod \"bbe812c3-c65e-4a52-a87e-7138189e359a\" (UID: \"bbe812c3-c65e-4a52-a87e-7138189e359a\") " Oct 07 13:56:27 crc kubenswrapper[4959]: I1007 13:56:27.194991 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bbe812c3-c65e-4a52-a87e-7138189e359a-bundle\") pod \"bbe812c3-c65e-4a52-a87e-7138189e359a\" (UID: \"bbe812c3-c65e-4a52-a87e-7138189e359a\") " Oct 07 13:56:27 crc kubenswrapper[4959]: I1007 13:56:27.195033 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bbe812c3-c65e-4a52-a87e-7138189e359a-util\") pod \"bbe812c3-c65e-4a52-a87e-7138189e359a\" (UID: \"bbe812c3-c65e-4a52-a87e-7138189e359a\") " Oct 07 13:56:27 crc kubenswrapper[4959]: I1007 13:56:27.196327 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbe812c3-c65e-4a52-a87e-7138189e359a-bundle" (OuterVolumeSpecName: "bundle") pod "bbe812c3-c65e-4a52-a87e-7138189e359a" (UID: "bbe812c3-c65e-4a52-a87e-7138189e359a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:56:27 crc kubenswrapper[4959]: I1007 13:56:27.205487 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbe812c3-c65e-4a52-a87e-7138189e359a-kube-api-access-zfqfc" (OuterVolumeSpecName: "kube-api-access-zfqfc") pod "bbe812c3-c65e-4a52-a87e-7138189e359a" (UID: "bbe812c3-c65e-4a52-a87e-7138189e359a"). InnerVolumeSpecName "kube-api-access-zfqfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:56:27 crc kubenswrapper[4959]: I1007 13:56:27.209456 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbe812c3-c65e-4a52-a87e-7138189e359a-util" (OuterVolumeSpecName: "util") pod "bbe812c3-c65e-4a52-a87e-7138189e359a" (UID: "bbe812c3-c65e-4a52-a87e-7138189e359a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:56:27 crc kubenswrapper[4959]: I1007 13:56:27.297015 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfqfc\" (UniqueName: \"kubernetes.io/projected/bbe812c3-c65e-4a52-a87e-7138189e359a-kube-api-access-zfqfc\") on node \"crc\" DevicePath \"\"" Oct 07 13:56:27 crc kubenswrapper[4959]: I1007 13:56:27.297060 4959 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bbe812c3-c65e-4a52-a87e-7138189e359a-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 13:56:27 crc kubenswrapper[4959]: I1007 13:56:27.297076 4959 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bbe812c3-c65e-4a52-a87e-7138189e359a-util\") on node \"crc\" DevicePath \"\"" Oct 07 13:56:27 crc kubenswrapper[4959]: I1007 13:56:27.793533 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz" event={"ID":"bbe812c3-c65e-4a52-a87e-7138189e359a","Type":"ContainerDied","Data":"c2446f43de05328ded1c14725115a69a3a0156d8e49fb8295ab0ddd7f078d89b"} Oct 07 13:56:27 crc kubenswrapper[4959]: I1007 13:56:27.793625 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2446f43de05328ded1c14725115a69a3a0156d8e49fb8295ab0ddd7f078d89b" Oct 07 13:56:27 crc kubenswrapper[4959]: I1007 13:56:27.794257 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz" Oct 07 13:56:31 crc kubenswrapper[4959]: I1007 13:56:31.914923 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-qf9tf"] Oct 07 13:56:31 crc kubenswrapper[4959]: E1007 13:56:31.916249 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbe812c3-c65e-4a52-a87e-7138189e359a" containerName="extract" Oct 07 13:56:31 crc kubenswrapper[4959]: I1007 13:56:31.916341 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbe812c3-c65e-4a52-a87e-7138189e359a" containerName="extract" Oct 07 13:56:31 crc kubenswrapper[4959]: E1007 13:56:31.916417 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbe812c3-c65e-4a52-a87e-7138189e359a" containerName="pull" Oct 07 13:56:31 crc kubenswrapper[4959]: I1007 13:56:31.916481 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbe812c3-c65e-4a52-a87e-7138189e359a" containerName="pull" Oct 07 13:56:31 crc kubenswrapper[4959]: E1007 13:56:31.916607 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbe812c3-c65e-4a52-a87e-7138189e359a" containerName="util" Oct 07 13:56:31 crc kubenswrapper[4959]: I1007 13:56:31.916680 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbe812c3-c65e-4a52-a87e-7138189e359a" containerName="util" Oct 07 13:56:31 crc kubenswrapper[4959]: I1007 13:56:31.916884 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbe812c3-c65e-4a52-a87e-7138189e359a" containerName="extract" Oct 07 13:56:31 crc kubenswrapper[4959]: I1007 13:56:31.917491 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-qf9tf" Oct 07 13:56:31 crc kubenswrapper[4959]: I1007 13:56:31.920380 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-k2nzs" Oct 07 13:56:31 crc kubenswrapper[4959]: I1007 13:56:31.920905 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 07 13:56:31 crc kubenswrapper[4959]: I1007 13:56:31.921188 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 07 13:56:31 crc kubenswrapper[4959]: I1007 13:56:31.928827 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-qf9tf"] Oct 07 13:56:31 crc kubenswrapper[4959]: I1007 13:56:31.978525 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8zwt\" (UniqueName: \"kubernetes.io/projected/9569c43f-7384-4561-b6fe-231eedaba53a-kube-api-access-g8zwt\") pod \"nmstate-operator-858ddd8f98-qf9tf\" (UID: \"9569c43f-7384-4561-b6fe-231eedaba53a\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-qf9tf" Oct 07 13:56:32 crc kubenswrapper[4959]: I1007 13:56:32.079988 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8zwt\" (UniqueName: \"kubernetes.io/projected/9569c43f-7384-4561-b6fe-231eedaba53a-kube-api-access-g8zwt\") pod \"nmstate-operator-858ddd8f98-qf9tf\" (UID: \"9569c43f-7384-4561-b6fe-231eedaba53a\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-qf9tf" Oct 07 13:56:32 crc kubenswrapper[4959]: I1007 13:56:32.110968 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8zwt\" (UniqueName: \"kubernetes.io/projected/9569c43f-7384-4561-b6fe-231eedaba53a-kube-api-access-g8zwt\") pod \"nmstate-operator-858ddd8f98-qf9tf\" (UID: \"9569c43f-7384-4561-b6fe-231eedaba53a\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-qf9tf" Oct 07 13:56:32 crc kubenswrapper[4959]: I1007 13:56:32.235885 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-qf9tf" Oct 07 13:56:32 crc kubenswrapper[4959]: I1007 13:56:32.503501 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-qf9tf"] Oct 07 13:56:32 crc kubenswrapper[4959]: I1007 13:56:32.821951 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-qf9tf" event={"ID":"9569c43f-7384-4561-b6fe-231eedaba53a","Type":"ContainerStarted","Data":"ac2dcc779e05da08f9fdf769320a644e9acec7b19579eaf55acc452f55ac278e"} Oct 07 13:56:34 crc kubenswrapper[4959]: I1007 13:56:34.839095 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-qf9tf" event={"ID":"9569c43f-7384-4561-b6fe-231eedaba53a","Type":"ContainerStarted","Data":"c755efb9baef88373646531ddbe5b0eb36453b8f479b1d0ec4e0558bf31ef6dd"} Oct 07 13:56:34 crc kubenswrapper[4959]: I1007 13:56:34.855373 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-qf9tf" podStartSLOduration=1.916942879 podStartE2EDuration="3.855346775s" podCreationTimestamp="2025-10-07 13:56:31 +0000 UTC" firstStartedPulling="2025-10-07 13:56:32.512851927 +0000 UTC m=+674.596256252" lastFinishedPulling="2025-10-07 13:56:34.451255813 +0000 UTC m=+676.534660148" observedRunningTime="2025-10-07 13:56:34.853445278 +0000 UTC m=+676.936849623" watchObservedRunningTime="2025-10-07 13:56:34.855346775 +0000 UTC m=+676.938751130" Oct 07 13:56:40 crc kubenswrapper[4959]: I1007 13:56:40.803178 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-f9jjx"] Oct 07 13:56:40 crc kubenswrapper[4959]: I1007 13:56:40.804426 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-f9jjx" Oct 07 13:56:40 crc kubenswrapper[4959]: I1007 13:56:40.808486 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-xft6j" Oct 07 13:56:40 crc kubenswrapper[4959]: I1007 13:56:40.815156 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-v5vcx"] Oct 07 13:56:40 crc kubenswrapper[4959]: I1007 13:56:40.815885 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-v5vcx" Oct 07 13:56:40 crc kubenswrapper[4959]: I1007 13:56:40.818587 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 07 13:56:40 crc kubenswrapper[4959]: I1007 13:56:40.843657 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-vsp7j"] Oct 07 13:56:40 crc kubenswrapper[4959]: I1007 13:56:40.845506 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-vsp7j" Oct 07 13:56:40 crc kubenswrapper[4959]: I1007 13:56:40.877388 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-f9jjx"] Oct 07 13:56:40 crc kubenswrapper[4959]: I1007 13:56:40.899597 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42bbc\" (UniqueName: \"kubernetes.io/projected/1e34329c-0703-41ea-81b5-cd2056ab1491-kube-api-access-42bbc\") pod \"nmstate-metrics-fdff9cb8d-f9jjx\" (UID: \"1e34329c-0703-41ea-81b5-cd2056ab1491\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-f9jjx" Oct 07 13:56:40 crc kubenswrapper[4959]: I1007 13:56:40.899652 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/013ff83c-f506-43cc-877b-dc67c6bb7c10-ovs-socket\") pod \"nmstate-handler-vsp7j\" (UID: \"013ff83c-f506-43cc-877b-dc67c6bb7c10\") " pod="openshift-nmstate/nmstate-handler-vsp7j" Oct 07 13:56:40 crc kubenswrapper[4959]: I1007 13:56:40.899695 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zzsx\" (UniqueName: \"kubernetes.io/projected/5d7ebbd8-5199-43e0-a92d-54f120ee968a-kube-api-access-4zzsx\") pod \"nmstate-webhook-6cdbc54649-v5vcx\" (UID: \"5d7ebbd8-5199-43e0-a92d-54f120ee968a\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-v5vcx" Oct 07 13:56:40 crc kubenswrapper[4959]: I1007 13:56:40.899714 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9hd6\" (UniqueName: \"kubernetes.io/projected/013ff83c-f506-43cc-877b-dc67c6bb7c10-kube-api-access-k9hd6\") pod \"nmstate-handler-vsp7j\" (UID: \"013ff83c-f506-43cc-877b-dc67c6bb7c10\") " pod="openshift-nmstate/nmstate-handler-vsp7j" Oct 07 13:56:40 crc kubenswrapper[4959]: I1007 13:56:40.899762 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/013ff83c-f506-43cc-877b-dc67c6bb7c10-nmstate-lock\") pod \"nmstate-handler-vsp7j\" (UID: \"013ff83c-f506-43cc-877b-dc67c6bb7c10\") " pod="openshift-nmstate/nmstate-handler-vsp7j" Oct 07 13:56:40 crc kubenswrapper[4959]: I1007 13:56:40.899785 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/013ff83c-f506-43cc-877b-dc67c6bb7c10-dbus-socket\") pod \"nmstate-handler-vsp7j\" (UID: \"013ff83c-f506-43cc-877b-dc67c6bb7c10\") " pod="openshift-nmstate/nmstate-handler-vsp7j" Oct 07 13:56:40 crc kubenswrapper[4959]: I1007 13:56:40.899801 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/5d7ebbd8-5199-43e0-a92d-54f120ee968a-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-v5vcx\" (UID: \"5d7ebbd8-5199-43e0-a92d-54f120ee968a\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-v5vcx" Oct 07 13:56:40 crc kubenswrapper[4959]: I1007 13:56:40.918995 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-v5vcx"] Oct 07 13:56:40 crc kubenswrapper[4959]: I1007 13:56:40.931655 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-t7bgt"] Oct 07 13:56:40 crc kubenswrapper[4959]: I1007 13:56:40.932470 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-t7bgt" Oct 07 13:56:40 crc kubenswrapper[4959]: I1007 13:56:40.935884 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-nqrnm" Oct 07 13:56:40 crc kubenswrapper[4959]: I1007 13:56:40.936297 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 07 13:56:40 crc kubenswrapper[4959]: I1007 13:56:40.936462 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 07 13:56:40 crc kubenswrapper[4959]: I1007 13:56:40.945884 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-t7bgt"] Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.000918 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/5d7ebbd8-5199-43e0-a92d-54f120ee968a-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-v5vcx\" (UID: \"5d7ebbd8-5199-43e0-a92d-54f120ee968a\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-v5vcx" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.000983 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42bbc\" (UniqueName: \"kubernetes.io/projected/1e34329c-0703-41ea-81b5-cd2056ab1491-kube-api-access-42bbc\") pod \"nmstate-metrics-fdff9cb8d-f9jjx\" (UID: \"1e34329c-0703-41ea-81b5-cd2056ab1491\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-f9jjx" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.001013 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/013ff83c-f506-43cc-877b-dc67c6bb7c10-ovs-socket\") pod \"nmstate-handler-vsp7j\" (UID: \"013ff83c-f506-43cc-877b-dc67c6bb7c10\") " pod="openshift-nmstate/nmstate-handler-vsp7j" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.001044 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/0d2a66c5-8366-46e5-9c16-ebf1833b0207-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-t7bgt\" (UID: \"0d2a66c5-8366-46e5-9c16-ebf1833b0207\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-t7bgt" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.001065 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zzsx\" (UniqueName: \"kubernetes.io/projected/5d7ebbd8-5199-43e0-a92d-54f120ee968a-kube-api-access-4zzsx\") pod \"nmstate-webhook-6cdbc54649-v5vcx\" (UID: \"5d7ebbd8-5199-43e0-a92d-54f120ee968a\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-v5vcx" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.001080 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9hd6\" (UniqueName: \"kubernetes.io/projected/013ff83c-f506-43cc-877b-dc67c6bb7c10-kube-api-access-k9hd6\") pod \"nmstate-handler-vsp7j\" (UID: \"013ff83c-f506-43cc-877b-dc67c6bb7c10\") " pod="openshift-nmstate/nmstate-handler-vsp7j" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.001104 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/0d2a66c5-8366-46e5-9c16-ebf1833b0207-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-t7bgt\" (UID: \"0d2a66c5-8366-46e5-9c16-ebf1833b0207\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-t7bgt" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.001178 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/013ff83c-f506-43cc-877b-dc67c6bb7c10-nmstate-lock\") pod \"nmstate-handler-vsp7j\" (UID: \"013ff83c-f506-43cc-877b-dc67c6bb7c10\") " pod="openshift-nmstate/nmstate-handler-vsp7j" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.001195 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/013ff83c-f506-43cc-877b-dc67c6bb7c10-dbus-socket\") pod \"nmstate-handler-vsp7j\" (UID: \"013ff83c-f506-43cc-877b-dc67c6bb7c10\") " pod="openshift-nmstate/nmstate-handler-vsp7j" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.001213 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zgfs\" (UniqueName: \"kubernetes.io/projected/0d2a66c5-8366-46e5-9c16-ebf1833b0207-kube-api-access-9zgfs\") pod \"nmstate-console-plugin-6b874cbd85-t7bgt\" (UID: \"0d2a66c5-8366-46e5-9c16-ebf1833b0207\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-t7bgt" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.001402 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/013ff83c-f506-43cc-877b-dc67c6bb7c10-ovs-socket\") pod \"nmstate-handler-vsp7j\" (UID: \"013ff83c-f506-43cc-877b-dc67c6bb7c10\") " pod="openshift-nmstate/nmstate-handler-vsp7j" Oct 07 13:56:41 crc kubenswrapper[4959]: E1007 13:56:41.001127 4959 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 07 13:56:41 crc kubenswrapper[4959]: E1007 13:56:41.001463 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d7ebbd8-5199-43e0-a92d-54f120ee968a-tls-key-pair podName:5d7ebbd8-5199-43e0-a92d-54f120ee968a nodeName:}" failed. No retries permitted until 2025-10-07 13:56:41.501447812 +0000 UTC m=+683.584852127 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/5d7ebbd8-5199-43e0-a92d-54f120ee968a-tls-key-pair") pod "nmstate-webhook-6cdbc54649-v5vcx" (UID: "5d7ebbd8-5199-43e0-a92d-54f120ee968a") : secret "openshift-nmstate-webhook" not found Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.001495 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/013ff83c-f506-43cc-877b-dc67c6bb7c10-nmstate-lock\") pod \"nmstate-handler-vsp7j\" (UID: \"013ff83c-f506-43cc-877b-dc67c6bb7c10\") " pod="openshift-nmstate/nmstate-handler-vsp7j" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.001753 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/013ff83c-f506-43cc-877b-dc67c6bb7c10-dbus-socket\") pod \"nmstate-handler-vsp7j\" (UID: \"013ff83c-f506-43cc-877b-dc67c6bb7c10\") " pod="openshift-nmstate/nmstate-handler-vsp7j" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.018491 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42bbc\" (UniqueName: \"kubernetes.io/projected/1e34329c-0703-41ea-81b5-cd2056ab1491-kube-api-access-42bbc\") pod \"nmstate-metrics-fdff9cb8d-f9jjx\" (UID: \"1e34329c-0703-41ea-81b5-cd2056ab1491\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-f9jjx" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.021181 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9hd6\" (UniqueName: \"kubernetes.io/projected/013ff83c-f506-43cc-877b-dc67c6bb7c10-kube-api-access-k9hd6\") pod \"nmstate-handler-vsp7j\" (UID: \"013ff83c-f506-43cc-877b-dc67c6bb7c10\") " pod="openshift-nmstate/nmstate-handler-vsp7j" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.022838 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zzsx\" (UniqueName: \"kubernetes.io/projected/5d7ebbd8-5199-43e0-a92d-54f120ee968a-kube-api-access-4zzsx\") pod \"nmstate-webhook-6cdbc54649-v5vcx\" (UID: \"5d7ebbd8-5199-43e0-a92d-54f120ee968a\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-v5vcx" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.102441 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zgfs\" (UniqueName: \"kubernetes.io/projected/0d2a66c5-8366-46e5-9c16-ebf1833b0207-kube-api-access-9zgfs\") pod \"nmstate-console-plugin-6b874cbd85-t7bgt\" (UID: \"0d2a66c5-8366-46e5-9c16-ebf1833b0207\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-t7bgt" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.103031 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/0d2a66c5-8366-46e5-9c16-ebf1833b0207-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-t7bgt\" (UID: \"0d2a66c5-8366-46e5-9c16-ebf1833b0207\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-t7bgt" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.103145 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/0d2a66c5-8366-46e5-9c16-ebf1833b0207-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-t7bgt\" (UID: \"0d2a66c5-8366-46e5-9c16-ebf1833b0207\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-t7bgt" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.104002 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/0d2a66c5-8366-46e5-9c16-ebf1833b0207-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-t7bgt\" (UID: \"0d2a66c5-8366-46e5-9c16-ebf1833b0207\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-t7bgt" Oct 07 13:56:41 crc kubenswrapper[4959]: E1007 13:56:41.104164 4959 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Oct 07 13:56:41 crc kubenswrapper[4959]: E1007 13:56:41.104281 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0d2a66c5-8366-46e5-9c16-ebf1833b0207-plugin-serving-cert podName:0d2a66c5-8366-46e5-9c16-ebf1833b0207 nodeName:}" failed. No retries permitted until 2025-10-07 13:56:41.604267049 +0000 UTC m=+683.687671374 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/0d2a66c5-8366-46e5-9c16-ebf1833b0207-plugin-serving-cert") pod "nmstate-console-plugin-6b874cbd85-t7bgt" (UID: "0d2a66c5-8366-46e5-9c16-ebf1833b0207") : secret "plugin-serving-cert" not found Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.114073 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-9ffd69db5-5tjjs"] Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.114936 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-9ffd69db5-5tjjs" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.129850 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zgfs\" (UniqueName: \"kubernetes.io/projected/0d2a66c5-8366-46e5-9c16-ebf1833b0207-kube-api-access-9zgfs\") pod \"nmstate-console-plugin-6b874cbd85-t7bgt\" (UID: \"0d2a66c5-8366-46e5-9c16-ebf1833b0207\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-t7bgt" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.132347 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-9ffd69db5-5tjjs"] Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.135728 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-f9jjx" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.167767 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-vsp7j" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.204804 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/87cd85f3-42b9-4923-a72c-9947cf4960b2-console-oauth-config\") pod \"console-9ffd69db5-5tjjs\" (UID: \"87cd85f3-42b9-4923-a72c-9947cf4960b2\") " pod="openshift-console/console-9ffd69db5-5tjjs" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.205190 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/87cd85f3-42b9-4923-a72c-9947cf4960b2-console-serving-cert\") pod \"console-9ffd69db5-5tjjs\" (UID: \"87cd85f3-42b9-4923-a72c-9947cf4960b2\") " pod="openshift-console/console-9ffd69db5-5tjjs" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.205241 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/87cd85f3-42b9-4923-a72c-9947cf4960b2-service-ca\") pod \"console-9ffd69db5-5tjjs\" (UID: \"87cd85f3-42b9-4923-a72c-9947cf4960b2\") " pod="openshift-console/console-9ffd69db5-5tjjs" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.205287 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/87cd85f3-42b9-4923-a72c-9947cf4960b2-console-config\") pod \"console-9ffd69db5-5tjjs\" (UID: \"87cd85f3-42b9-4923-a72c-9947cf4960b2\") " pod="openshift-console/console-9ffd69db5-5tjjs" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.205314 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/87cd85f3-42b9-4923-a72c-9947cf4960b2-trusted-ca-bundle\") pod \"console-9ffd69db5-5tjjs\" (UID: \"87cd85f3-42b9-4923-a72c-9947cf4960b2\") " pod="openshift-console/console-9ffd69db5-5tjjs" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.205345 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84lr2\" (UniqueName: \"kubernetes.io/projected/87cd85f3-42b9-4923-a72c-9947cf4960b2-kube-api-access-84lr2\") pod \"console-9ffd69db5-5tjjs\" (UID: \"87cd85f3-42b9-4923-a72c-9947cf4960b2\") " pod="openshift-console/console-9ffd69db5-5tjjs" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.205400 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/87cd85f3-42b9-4923-a72c-9947cf4960b2-oauth-serving-cert\") pod \"console-9ffd69db5-5tjjs\" (UID: \"87cd85f3-42b9-4923-a72c-9947cf4960b2\") " pod="openshift-console/console-9ffd69db5-5tjjs" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.308700 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/87cd85f3-42b9-4923-a72c-9947cf4960b2-trusted-ca-bundle\") pod \"console-9ffd69db5-5tjjs\" (UID: \"87cd85f3-42b9-4923-a72c-9947cf4960b2\") " pod="openshift-console/console-9ffd69db5-5tjjs" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.308741 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/87cd85f3-42b9-4923-a72c-9947cf4960b2-console-config\") pod \"console-9ffd69db5-5tjjs\" (UID: \"87cd85f3-42b9-4923-a72c-9947cf4960b2\") " pod="openshift-console/console-9ffd69db5-5tjjs" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.308765 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84lr2\" (UniqueName: \"kubernetes.io/projected/87cd85f3-42b9-4923-a72c-9947cf4960b2-kube-api-access-84lr2\") pod \"console-9ffd69db5-5tjjs\" (UID: \"87cd85f3-42b9-4923-a72c-9947cf4960b2\") " pod="openshift-console/console-9ffd69db5-5tjjs" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.308812 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/87cd85f3-42b9-4923-a72c-9947cf4960b2-oauth-serving-cert\") pod \"console-9ffd69db5-5tjjs\" (UID: \"87cd85f3-42b9-4923-a72c-9947cf4960b2\") " pod="openshift-console/console-9ffd69db5-5tjjs" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.308830 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/87cd85f3-42b9-4923-a72c-9947cf4960b2-console-oauth-config\") pod \"console-9ffd69db5-5tjjs\" (UID: \"87cd85f3-42b9-4923-a72c-9947cf4960b2\") " pod="openshift-console/console-9ffd69db5-5tjjs" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.308860 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/87cd85f3-42b9-4923-a72c-9947cf4960b2-console-serving-cert\") pod \"console-9ffd69db5-5tjjs\" (UID: \"87cd85f3-42b9-4923-a72c-9947cf4960b2\") " pod="openshift-console/console-9ffd69db5-5tjjs" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.308883 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/87cd85f3-42b9-4923-a72c-9947cf4960b2-service-ca\") pod \"console-9ffd69db5-5tjjs\" (UID: \"87cd85f3-42b9-4923-a72c-9947cf4960b2\") " pod="openshift-console/console-9ffd69db5-5tjjs" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.309669 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/87cd85f3-42b9-4923-a72c-9947cf4960b2-service-ca\") pod \"console-9ffd69db5-5tjjs\" (UID: \"87cd85f3-42b9-4923-a72c-9947cf4960b2\") " pod="openshift-console/console-9ffd69db5-5tjjs" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.311668 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/87cd85f3-42b9-4923-a72c-9947cf4960b2-oauth-serving-cert\") pod \"console-9ffd69db5-5tjjs\" (UID: \"87cd85f3-42b9-4923-a72c-9947cf4960b2\") " pod="openshift-console/console-9ffd69db5-5tjjs" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.312793 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/87cd85f3-42b9-4923-a72c-9947cf4960b2-console-config\") pod \"console-9ffd69db5-5tjjs\" (UID: \"87cd85f3-42b9-4923-a72c-9947cf4960b2\") " pod="openshift-console/console-9ffd69db5-5tjjs" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.314084 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/87cd85f3-42b9-4923-a72c-9947cf4960b2-console-oauth-config\") pod \"console-9ffd69db5-5tjjs\" (UID: \"87cd85f3-42b9-4923-a72c-9947cf4960b2\") " pod="openshift-console/console-9ffd69db5-5tjjs" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.315504 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/87cd85f3-42b9-4923-a72c-9947cf4960b2-trusted-ca-bundle\") pod \"console-9ffd69db5-5tjjs\" (UID: \"87cd85f3-42b9-4923-a72c-9947cf4960b2\") " pod="openshift-console/console-9ffd69db5-5tjjs" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.317811 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/87cd85f3-42b9-4923-a72c-9947cf4960b2-console-serving-cert\") pod \"console-9ffd69db5-5tjjs\" (UID: \"87cd85f3-42b9-4923-a72c-9947cf4960b2\") " pod="openshift-console/console-9ffd69db5-5tjjs" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.324912 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-f9jjx"] Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.332190 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84lr2\" (UniqueName: \"kubernetes.io/projected/87cd85f3-42b9-4923-a72c-9947cf4960b2-kube-api-access-84lr2\") pod \"console-9ffd69db5-5tjjs\" (UID: \"87cd85f3-42b9-4923-a72c-9947cf4960b2\") " pod="openshift-console/console-9ffd69db5-5tjjs" Oct 07 13:56:41 crc kubenswrapper[4959]: W1007 13:56:41.333345 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e34329c_0703_41ea_81b5_cd2056ab1491.slice/crio-738043fdb1b4dde2a1f455c20c2d836e833d80b73307f3ffe5c45509d61a7454 WatchSource:0}: Error finding container 738043fdb1b4dde2a1f455c20c2d836e833d80b73307f3ffe5c45509d61a7454: Status 404 returned error can't find the container with id 738043fdb1b4dde2a1f455c20c2d836e833d80b73307f3ffe5c45509d61a7454 Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.455856 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-9ffd69db5-5tjjs" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.511811 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/5d7ebbd8-5199-43e0-a92d-54f120ee968a-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-v5vcx\" (UID: \"5d7ebbd8-5199-43e0-a92d-54f120ee968a\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-v5vcx" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.516821 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/5d7ebbd8-5199-43e0-a92d-54f120ee968a-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-v5vcx\" (UID: \"5d7ebbd8-5199-43e0-a92d-54f120ee968a\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-v5vcx" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.613650 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/0d2a66c5-8366-46e5-9c16-ebf1833b0207-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-t7bgt\" (UID: \"0d2a66c5-8366-46e5-9c16-ebf1833b0207\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-t7bgt" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.619979 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/0d2a66c5-8366-46e5-9c16-ebf1833b0207-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-t7bgt\" (UID: \"0d2a66c5-8366-46e5-9c16-ebf1833b0207\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-t7bgt" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.668159 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-9ffd69db5-5tjjs"] Oct 07 13:56:41 crc kubenswrapper[4959]: W1007 13:56:41.672168 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87cd85f3_42b9_4923_a72c_9947cf4960b2.slice/crio-b4ca4f2698db884d8fdd6b480a3871c7f83df721b3908947ec77d71f700c5c68 WatchSource:0}: Error finding container b4ca4f2698db884d8fdd6b480a3871c7f83df721b3908947ec77d71f700c5c68: Status 404 returned error can't find the container with id b4ca4f2698db884d8fdd6b480a3871c7f83df721b3908947ec77d71f700c5c68 Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.755493 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-v5vcx" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.847183 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-t7bgt" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.877204 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-vsp7j" event={"ID":"013ff83c-f506-43cc-877b-dc67c6bb7c10","Type":"ContainerStarted","Data":"88058754fce86ce507538e305938c356667bb7702274476c1f96e514f04cb6b8"} Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.879195 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-9ffd69db5-5tjjs" event={"ID":"87cd85f3-42b9-4923-a72c-9947cf4960b2","Type":"ContainerStarted","Data":"9fd672ef62f348f5a771f9ec0549a6605c5b1da5da8abb9bacf3af05298a062f"} Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.879228 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-9ffd69db5-5tjjs" event={"ID":"87cd85f3-42b9-4923-a72c-9947cf4960b2","Type":"ContainerStarted","Data":"b4ca4f2698db884d8fdd6b480a3871c7f83df721b3908947ec77d71f700c5c68"} Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.880698 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-f9jjx" event={"ID":"1e34329c-0703-41ea-81b5-cd2056ab1491","Type":"ContainerStarted","Data":"738043fdb1b4dde2a1f455c20c2d836e833d80b73307f3ffe5c45509d61a7454"} Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.903584 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-9ffd69db5-5tjjs" podStartSLOduration=0.903536031 podStartE2EDuration="903.536031ms" podCreationTimestamp="2025-10-07 13:56:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:56:41.895783159 +0000 UTC m=+683.979187564" watchObservedRunningTime="2025-10-07 13:56:41.903536031 +0000 UTC m=+683.986940366" Oct 07 13:56:41 crc kubenswrapper[4959]: I1007 13:56:41.990019 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-v5vcx"] Oct 07 13:56:42 crc kubenswrapper[4959]: W1007 13:56:42.003046 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d7ebbd8_5199_43e0_a92d_54f120ee968a.slice/crio-4700e9b6b5ea4ebe4da3f16817c97f7338a383233193a9c91b54c52133563872 WatchSource:0}: Error finding container 4700e9b6b5ea4ebe4da3f16817c97f7338a383233193a9c91b54c52133563872: Status 404 returned error can't find the container with id 4700e9b6b5ea4ebe4da3f16817c97f7338a383233193a9c91b54c52133563872 Oct 07 13:56:42 crc kubenswrapper[4959]: I1007 13:56:42.114369 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-t7bgt"] Oct 07 13:56:42 crc kubenswrapper[4959]: W1007 13:56:42.118697 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0d2a66c5_8366_46e5_9c16_ebf1833b0207.slice/crio-05b171ac99b58955d244ba567b4c9e70839ee91a4f5ae09942b6034440cecc31 WatchSource:0}: Error finding container 05b171ac99b58955d244ba567b4c9e70839ee91a4f5ae09942b6034440cecc31: Status 404 returned error can't find the container with id 05b171ac99b58955d244ba567b4c9e70839ee91a4f5ae09942b6034440cecc31 Oct 07 13:56:42 crc kubenswrapper[4959]: I1007 13:56:42.888780 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-t7bgt" event={"ID":"0d2a66c5-8366-46e5-9c16-ebf1833b0207","Type":"ContainerStarted","Data":"05b171ac99b58955d244ba567b4c9e70839ee91a4f5ae09942b6034440cecc31"} Oct 07 13:56:42 crc kubenswrapper[4959]: I1007 13:56:42.890500 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-v5vcx" event={"ID":"5d7ebbd8-5199-43e0-a92d-54f120ee968a","Type":"ContainerStarted","Data":"4700e9b6b5ea4ebe4da3f16817c97f7338a383233193a9c91b54c52133563872"} Oct 07 13:56:51 crc kubenswrapper[4959]: I1007 13:56:51.457049 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-9ffd69db5-5tjjs" Oct 07 13:56:51 crc kubenswrapper[4959]: I1007 13:56:51.457960 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-9ffd69db5-5tjjs" Oct 07 13:56:51 crc kubenswrapper[4959]: I1007 13:56:51.462068 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-9ffd69db5-5tjjs" Oct 07 13:56:51 crc kubenswrapper[4959]: I1007 13:56:51.955542 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-f9jjx" event={"ID":"1e34329c-0703-41ea-81b5-cd2056ab1491","Type":"ContainerStarted","Data":"d05bba42f8bdeeed365d6bda8cb424ccc627318c14f2e8d365b1b970c39df8d6"} Oct 07 13:56:51 crc kubenswrapper[4959]: I1007 13:56:51.957456 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-v5vcx" event={"ID":"5d7ebbd8-5199-43e0-a92d-54f120ee968a","Type":"ContainerStarted","Data":"92cfa4e2d828dfc6c35a4209469227d493827876e87deba15a2bfd7d77db3f00"} Oct 07 13:56:51 crc kubenswrapper[4959]: I1007 13:56:51.957582 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-v5vcx" Oct 07 13:56:51 crc kubenswrapper[4959]: I1007 13:56:51.959088 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-vsp7j" event={"ID":"013ff83c-f506-43cc-877b-dc67c6bb7c10","Type":"ContainerStarted","Data":"2443e75593dfc3eaceba6de3707a51b057f49af34890aefc0560c4edbab45606"} Oct 07 13:56:51 crc kubenswrapper[4959]: I1007 13:56:51.963949 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-9ffd69db5-5tjjs" Oct 07 13:56:51 crc kubenswrapper[4959]: I1007 13:56:51.999764 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-v5vcx" podStartSLOduration=3.384027763 podStartE2EDuration="11.999747838s" podCreationTimestamp="2025-10-07 13:56:40 +0000 UTC" firstStartedPulling="2025-10-07 13:56:42.005943679 +0000 UTC m=+684.089347994" lastFinishedPulling="2025-10-07 13:56:50.621663734 +0000 UTC m=+692.705068069" observedRunningTime="2025-10-07 13:56:51.97844477 +0000 UTC m=+694.061849095" watchObservedRunningTime="2025-10-07 13:56:51.999747838 +0000 UTC m=+694.083152163" Oct 07 13:56:52 crc kubenswrapper[4959]: I1007 13:56:52.001398 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-vsp7j" podStartSLOduration=2.571871799 podStartE2EDuration="12.001391598s" podCreationTimestamp="2025-10-07 13:56:40 +0000 UTC" firstStartedPulling="2025-10-07 13:56:41.188798543 +0000 UTC m=+683.272202868" lastFinishedPulling="2025-10-07 13:56:50.618318302 +0000 UTC m=+692.701722667" observedRunningTime="2025-10-07 13:56:51.997080962 +0000 UTC m=+694.080485287" watchObservedRunningTime="2025-10-07 13:56:52.001391598 +0000 UTC m=+694.084795923" Oct 07 13:56:52 crc kubenswrapper[4959]: I1007 13:56:52.050143 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-pcvlp"] Oct 07 13:56:52 crc kubenswrapper[4959]: I1007 13:56:52.972808 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-vsp7j" Oct 07 13:56:53 crc kubenswrapper[4959]: I1007 13:56:53.979423 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-t7bgt" event={"ID":"0d2a66c5-8366-46e5-9c16-ebf1833b0207","Type":"ContainerStarted","Data":"5a8b46a2e9fc308f122c83d03718012de96d03397fe2e83cdeb60064d92b99cf"} Oct 07 13:56:53 crc kubenswrapper[4959]: I1007 13:56:53.999135 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-t7bgt" podStartSLOduration=3.286960229 podStartE2EDuration="13.999090445s" podCreationTimestamp="2025-10-07 13:56:40 +0000 UTC" firstStartedPulling="2025-10-07 13:56:42.120720633 +0000 UTC m=+684.204124958" lastFinishedPulling="2025-10-07 13:56:52.832850839 +0000 UTC m=+694.916255174" observedRunningTime="2025-10-07 13:56:53.99323836 +0000 UTC m=+696.076642685" watchObservedRunningTime="2025-10-07 13:56:53.999090445 +0000 UTC m=+696.082494810" Oct 07 13:56:56 crc kubenswrapper[4959]: I1007 13:56:56.206621 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-vsp7j" Oct 07 13:56:57 crc kubenswrapper[4959]: I1007 13:56:57.139485 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-f9jjx" event={"ID":"1e34329c-0703-41ea-81b5-cd2056ab1491","Type":"ContainerStarted","Data":"5096f9a6ab608d56c6c14b9b0f3f6106bd0835de35363032c26f9f57895d9684"} Oct 07 13:56:57 crc kubenswrapper[4959]: I1007 13:56:57.157174 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-f9jjx" podStartSLOduration=2.695919115 podStartE2EDuration="17.15715194s" podCreationTimestamp="2025-10-07 13:56:40 +0000 UTC" firstStartedPulling="2025-10-07 13:56:41.335399905 +0000 UTC m=+683.418804230" lastFinishedPulling="2025-10-07 13:56:55.79663269 +0000 UTC m=+697.880037055" observedRunningTime="2025-10-07 13:56:57.153081479 +0000 UTC m=+699.236485814" watchObservedRunningTime="2025-10-07 13:56:57.15715194 +0000 UTC m=+699.240556265" Oct 07 13:57:01 crc kubenswrapper[4959]: I1007 13:57:01.764295 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-v5vcx" Oct 07 13:57:13 crc kubenswrapper[4959]: I1007 13:57:13.981628 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c"] Oct 07 13:57:13 crc kubenswrapper[4959]: I1007 13:57:13.983625 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c" Oct 07 13:57:13 crc kubenswrapper[4959]: I1007 13:57:13.985392 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 07 13:57:13 crc kubenswrapper[4959]: I1007 13:57:13.997923 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c"] Oct 07 13:57:14 crc kubenswrapper[4959]: I1007 13:57:14.105609 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b524d641-ed92-49cc-b5f1-ff1d8addce42-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c\" (UID: \"b524d641-ed92-49cc-b5f1-ff1d8addce42\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c" Oct 07 13:57:14 crc kubenswrapper[4959]: I1007 13:57:14.105667 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b524d641-ed92-49cc-b5f1-ff1d8addce42-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c\" (UID: \"b524d641-ed92-49cc-b5f1-ff1d8addce42\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c" Oct 07 13:57:14 crc kubenswrapper[4959]: I1007 13:57:14.105737 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzkgf\" (UniqueName: \"kubernetes.io/projected/b524d641-ed92-49cc-b5f1-ff1d8addce42-kube-api-access-tzkgf\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c\" (UID: \"b524d641-ed92-49cc-b5f1-ff1d8addce42\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c" Oct 07 13:57:14 crc kubenswrapper[4959]: I1007 13:57:14.207178 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b524d641-ed92-49cc-b5f1-ff1d8addce42-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c\" (UID: \"b524d641-ed92-49cc-b5f1-ff1d8addce42\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c" Oct 07 13:57:14 crc kubenswrapper[4959]: I1007 13:57:14.207225 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b524d641-ed92-49cc-b5f1-ff1d8addce42-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c\" (UID: \"b524d641-ed92-49cc-b5f1-ff1d8addce42\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c" Oct 07 13:57:14 crc kubenswrapper[4959]: I1007 13:57:14.207275 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzkgf\" (UniqueName: \"kubernetes.io/projected/b524d641-ed92-49cc-b5f1-ff1d8addce42-kube-api-access-tzkgf\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c\" (UID: \"b524d641-ed92-49cc-b5f1-ff1d8addce42\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c" Oct 07 13:57:14 crc kubenswrapper[4959]: I1007 13:57:14.207700 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b524d641-ed92-49cc-b5f1-ff1d8addce42-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c\" (UID: \"b524d641-ed92-49cc-b5f1-ff1d8addce42\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c" Oct 07 13:57:14 crc kubenswrapper[4959]: I1007 13:57:14.207735 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b524d641-ed92-49cc-b5f1-ff1d8addce42-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c\" (UID: \"b524d641-ed92-49cc-b5f1-ff1d8addce42\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c" Oct 07 13:57:14 crc kubenswrapper[4959]: I1007 13:57:14.225955 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzkgf\" (UniqueName: \"kubernetes.io/projected/b524d641-ed92-49cc-b5f1-ff1d8addce42-kube-api-access-tzkgf\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c\" (UID: \"b524d641-ed92-49cc-b5f1-ff1d8addce42\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c" Oct 07 13:57:14 crc kubenswrapper[4959]: I1007 13:57:14.297765 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c" Oct 07 13:57:14 crc kubenswrapper[4959]: I1007 13:57:14.526915 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c"] Oct 07 13:57:14 crc kubenswrapper[4959]: W1007 13:57:14.533603 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb524d641_ed92_49cc_b5f1_ff1d8addce42.slice/crio-da78232322ecf44439ae20615b5d757cd46f06bc98580a610698f3c064cfb83d WatchSource:0}: Error finding container da78232322ecf44439ae20615b5d757cd46f06bc98580a610698f3c064cfb83d: Status 404 returned error can't find the container with id da78232322ecf44439ae20615b5d757cd46f06bc98580a610698f3c064cfb83d Oct 07 13:57:15 crc kubenswrapper[4959]: I1007 13:57:15.262219 4959 generic.go:334] "Generic (PLEG): container finished" podID="b524d641-ed92-49cc-b5f1-ff1d8addce42" containerID="1cee7d721d0707f325f3f6ca9ed19f10e40ca18862e313102471d4a6bded7125" exitCode=0 Oct 07 13:57:15 crc kubenswrapper[4959]: I1007 13:57:15.262260 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c" event={"ID":"b524d641-ed92-49cc-b5f1-ff1d8addce42","Type":"ContainerDied","Data":"1cee7d721d0707f325f3f6ca9ed19f10e40ca18862e313102471d4a6bded7125"} Oct 07 13:57:15 crc kubenswrapper[4959]: I1007 13:57:15.262302 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c" event={"ID":"b524d641-ed92-49cc-b5f1-ff1d8addce42","Type":"ContainerStarted","Data":"da78232322ecf44439ae20615b5d757cd46f06bc98580a610698f3c064cfb83d"} Oct 07 13:57:17 crc kubenswrapper[4959]: I1007 13:57:17.099401 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-pcvlp" podUID="4fca2a0e-545f-43b0-b4ec-e5f15babcd71" containerName="console" containerID="cri-o://d15871c5e854d27c40c25e1eca507e49d5a98e2edd4342505bbcf6c1517412fd" gracePeriod=15 Oct 07 13:57:17 crc kubenswrapper[4959]: I1007 13:57:17.275413 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-pcvlp_4fca2a0e-545f-43b0-b4ec-e5f15babcd71/console/0.log" Oct 07 13:57:17 crc kubenswrapper[4959]: I1007 13:57:17.275465 4959 generic.go:334] "Generic (PLEG): container finished" podID="4fca2a0e-545f-43b0-b4ec-e5f15babcd71" containerID="d15871c5e854d27c40c25e1eca507e49d5a98e2edd4342505bbcf6c1517412fd" exitCode=2 Oct 07 13:57:17 crc kubenswrapper[4959]: I1007 13:57:17.275498 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-pcvlp" event={"ID":"4fca2a0e-545f-43b0-b4ec-e5f15babcd71","Type":"ContainerDied","Data":"d15871c5e854d27c40c25e1eca507e49d5a98e2edd4342505bbcf6c1517412fd"} Oct 07 13:57:17 crc kubenswrapper[4959]: I1007 13:57:17.507768 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-pcvlp_4fca2a0e-545f-43b0-b4ec-e5f15babcd71/console/0.log" Oct 07 13:57:17 crc kubenswrapper[4959]: I1007 13:57:17.508308 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-pcvlp" Oct 07 13:57:17 crc kubenswrapper[4959]: I1007 13:57:17.553545 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-oauth-serving-cert\") pod \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\" (UID: \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\") " Oct 07 13:57:17 crc kubenswrapper[4959]: I1007 13:57:17.553595 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-console-config\") pod \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\" (UID: \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\") " Oct 07 13:57:17 crc kubenswrapper[4959]: I1007 13:57:17.553682 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-console-serving-cert\") pod \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\" (UID: \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\") " Oct 07 13:57:17 crc kubenswrapper[4959]: I1007 13:57:17.553716 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-console-oauth-config\") pod \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\" (UID: \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\") " Oct 07 13:57:17 crc kubenswrapper[4959]: I1007 13:57:17.553733 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxszd\" (UniqueName: \"kubernetes.io/projected/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-kube-api-access-qxszd\") pod \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\" (UID: \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\") " Oct 07 13:57:17 crc kubenswrapper[4959]: I1007 13:57:17.553775 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-service-ca\") pod \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\" (UID: \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\") " Oct 07 13:57:17 crc kubenswrapper[4959]: I1007 13:57:17.553790 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-trusted-ca-bundle\") pod \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\" (UID: \"4fca2a0e-545f-43b0-b4ec-e5f15babcd71\") " Oct 07 13:57:17 crc kubenswrapper[4959]: I1007 13:57:17.554911 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "4fca2a0e-545f-43b0-b4ec-e5f15babcd71" (UID: "4fca2a0e-545f-43b0-b4ec-e5f15babcd71"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:57:17 crc kubenswrapper[4959]: I1007 13:57:17.554923 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "4fca2a0e-545f-43b0-b4ec-e5f15babcd71" (UID: "4fca2a0e-545f-43b0-b4ec-e5f15babcd71"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:57:17 crc kubenswrapper[4959]: I1007 13:57:17.554940 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-console-config" (OuterVolumeSpecName: "console-config") pod "4fca2a0e-545f-43b0-b4ec-e5f15babcd71" (UID: "4fca2a0e-545f-43b0-b4ec-e5f15babcd71"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:57:17 crc kubenswrapper[4959]: I1007 13:57:17.555440 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-service-ca" (OuterVolumeSpecName: "service-ca") pod "4fca2a0e-545f-43b0-b4ec-e5f15babcd71" (UID: "4fca2a0e-545f-43b0-b4ec-e5f15babcd71"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:57:17 crc kubenswrapper[4959]: I1007 13:57:17.560366 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-kube-api-access-qxszd" (OuterVolumeSpecName: "kube-api-access-qxszd") pod "4fca2a0e-545f-43b0-b4ec-e5f15babcd71" (UID: "4fca2a0e-545f-43b0-b4ec-e5f15babcd71"). InnerVolumeSpecName "kube-api-access-qxszd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:57:17 crc kubenswrapper[4959]: I1007 13:57:17.560453 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "4fca2a0e-545f-43b0-b4ec-e5f15babcd71" (UID: "4fca2a0e-545f-43b0-b4ec-e5f15babcd71"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:57:17 crc kubenswrapper[4959]: I1007 13:57:17.560659 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "4fca2a0e-545f-43b0-b4ec-e5f15babcd71" (UID: "4fca2a0e-545f-43b0-b4ec-e5f15babcd71"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:57:17 crc kubenswrapper[4959]: I1007 13:57:17.655021 4959 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:57:17 crc kubenswrapper[4959]: I1007 13:57:17.655058 4959 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-console-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:57:17 crc kubenswrapper[4959]: I1007 13:57:17.655072 4959 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:57:17 crc kubenswrapper[4959]: I1007 13:57:17.655084 4959 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:57:17 crc kubenswrapper[4959]: I1007 13:57:17.655116 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxszd\" (UniqueName: \"kubernetes.io/projected/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-kube-api-access-qxszd\") on node \"crc\" DevicePath \"\"" Oct 07 13:57:17 crc kubenswrapper[4959]: I1007 13:57:17.655134 4959 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-service-ca\") on node \"crc\" DevicePath \"\"" Oct 07 13:57:17 crc kubenswrapper[4959]: I1007 13:57:17.655148 4959 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4fca2a0e-545f-43b0-b4ec-e5f15babcd71-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 13:57:18 crc kubenswrapper[4959]: I1007 13:57:18.285266 4959 generic.go:334] "Generic (PLEG): container finished" podID="b524d641-ed92-49cc-b5f1-ff1d8addce42" containerID="73b23f15b74dec60ac90cabd4695ef5e6a17fe8e1b4f4323da02a47b36d771b2" exitCode=0 Oct 07 13:57:18 crc kubenswrapper[4959]: I1007 13:57:18.285720 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c" event={"ID":"b524d641-ed92-49cc-b5f1-ff1d8addce42","Type":"ContainerDied","Data":"73b23f15b74dec60ac90cabd4695ef5e6a17fe8e1b4f4323da02a47b36d771b2"} Oct 07 13:57:18 crc kubenswrapper[4959]: I1007 13:57:18.288270 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-pcvlp_4fca2a0e-545f-43b0-b4ec-e5f15babcd71/console/0.log" Oct 07 13:57:18 crc kubenswrapper[4959]: I1007 13:57:18.288320 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-pcvlp" event={"ID":"4fca2a0e-545f-43b0-b4ec-e5f15babcd71","Type":"ContainerDied","Data":"56a8e024537dcef4e292424b730f2cf4efd6fb405e8b6b483bf8262bf5e4f1d4"} Oct 07 13:57:18 crc kubenswrapper[4959]: I1007 13:57:18.288356 4959 scope.go:117] "RemoveContainer" containerID="d15871c5e854d27c40c25e1eca507e49d5a98e2edd4342505bbcf6c1517412fd" Oct 07 13:57:18 crc kubenswrapper[4959]: I1007 13:57:18.288482 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-pcvlp" Oct 07 13:57:18 crc kubenswrapper[4959]: I1007 13:57:18.328391 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-pcvlp"] Oct 07 13:57:18 crc kubenswrapper[4959]: I1007 13:57:18.332219 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-pcvlp"] Oct 07 13:57:18 crc kubenswrapper[4959]: I1007 13:57:18.660851 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fca2a0e-545f-43b0-b4ec-e5f15babcd71" path="/var/lib/kubelet/pods/4fca2a0e-545f-43b0-b4ec-e5f15babcd71/volumes" Oct 07 13:57:19 crc kubenswrapper[4959]: I1007 13:57:19.299400 4959 generic.go:334] "Generic (PLEG): container finished" podID="b524d641-ed92-49cc-b5f1-ff1d8addce42" containerID="6f65a4ab6bb70dbbdbf1e6d49b9396dbf34a5a033766e3b2f657b75fa177a36e" exitCode=0 Oct 07 13:57:19 crc kubenswrapper[4959]: I1007 13:57:19.299507 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c" event={"ID":"b524d641-ed92-49cc-b5f1-ff1d8addce42","Type":"ContainerDied","Data":"6f65a4ab6bb70dbbdbf1e6d49b9396dbf34a5a033766e3b2f657b75fa177a36e"} Oct 07 13:57:20 crc kubenswrapper[4959]: I1007 13:57:20.629836 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 13:57:20 crc kubenswrapper[4959]: I1007 13:57:20.629921 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 13:57:20 crc kubenswrapper[4959]: I1007 13:57:20.646390 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c" Oct 07 13:57:20 crc kubenswrapper[4959]: I1007 13:57:20.703436 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b524d641-ed92-49cc-b5f1-ff1d8addce42-util\") pod \"b524d641-ed92-49cc-b5f1-ff1d8addce42\" (UID: \"b524d641-ed92-49cc-b5f1-ff1d8addce42\") " Oct 07 13:57:20 crc kubenswrapper[4959]: I1007 13:57:20.703539 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzkgf\" (UniqueName: \"kubernetes.io/projected/b524d641-ed92-49cc-b5f1-ff1d8addce42-kube-api-access-tzkgf\") pod \"b524d641-ed92-49cc-b5f1-ff1d8addce42\" (UID: \"b524d641-ed92-49cc-b5f1-ff1d8addce42\") " Oct 07 13:57:20 crc kubenswrapper[4959]: I1007 13:57:20.703612 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b524d641-ed92-49cc-b5f1-ff1d8addce42-bundle\") pod \"b524d641-ed92-49cc-b5f1-ff1d8addce42\" (UID: \"b524d641-ed92-49cc-b5f1-ff1d8addce42\") " Oct 07 13:57:20 crc kubenswrapper[4959]: I1007 13:57:20.705523 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b524d641-ed92-49cc-b5f1-ff1d8addce42-bundle" (OuterVolumeSpecName: "bundle") pod "b524d641-ed92-49cc-b5f1-ff1d8addce42" (UID: "b524d641-ed92-49cc-b5f1-ff1d8addce42"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:57:20 crc kubenswrapper[4959]: I1007 13:57:20.710320 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b524d641-ed92-49cc-b5f1-ff1d8addce42-kube-api-access-tzkgf" (OuterVolumeSpecName: "kube-api-access-tzkgf") pod "b524d641-ed92-49cc-b5f1-ff1d8addce42" (UID: "b524d641-ed92-49cc-b5f1-ff1d8addce42"). InnerVolumeSpecName "kube-api-access-tzkgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:57:20 crc kubenswrapper[4959]: I1007 13:57:20.777189 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b524d641-ed92-49cc-b5f1-ff1d8addce42-util" (OuterVolumeSpecName: "util") pod "b524d641-ed92-49cc-b5f1-ff1d8addce42" (UID: "b524d641-ed92-49cc-b5f1-ff1d8addce42"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:57:20 crc kubenswrapper[4959]: I1007 13:57:20.805519 4959 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b524d641-ed92-49cc-b5f1-ff1d8addce42-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 13:57:20 crc kubenswrapper[4959]: I1007 13:57:20.805554 4959 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b524d641-ed92-49cc-b5f1-ff1d8addce42-util\") on node \"crc\" DevicePath \"\"" Oct 07 13:57:20 crc kubenswrapper[4959]: I1007 13:57:20.805572 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzkgf\" (UniqueName: \"kubernetes.io/projected/b524d641-ed92-49cc-b5f1-ff1d8addce42-kube-api-access-tzkgf\") on node \"crc\" DevicePath \"\"" Oct 07 13:57:21 crc kubenswrapper[4959]: I1007 13:57:21.326406 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c" event={"ID":"b524d641-ed92-49cc-b5f1-ff1d8addce42","Type":"ContainerDied","Data":"da78232322ecf44439ae20615b5d757cd46f06bc98580a610698f3c064cfb83d"} Oct 07 13:57:21 crc kubenswrapper[4959]: I1007 13:57:21.326483 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c" Oct 07 13:57:21 crc kubenswrapper[4959]: I1007 13:57:21.326491 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da78232322ecf44439ae20615b5d757cd46f06bc98580a610698f3c064cfb83d" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.393293 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-9c45bf8cc-m9pg5"] Oct 07 13:57:29 crc kubenswrapper[4959]: E1007 13:57:29.393862 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b524d641-ed92-49cc-b5f1-ff1d8addce42" containerName="util" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.393875 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b524d641-ed92-49cc-b5f1-ff1d8addce42" containerName="util" Oct 07 13:57:29 crc kubenswrapper[4959]: E1007 13:57:29.393886 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b524d641-ed92-49cc-b5f1-ff1d8addce42" containerName="extract" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.393892 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b524d641-ed92-49cc-b5f1-ff1d8addce42" containerName="extract" Oct 07 13:57:29 crc kubenswrapper[4959]: E1007 13:57:29.393904 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b524d641-ed92-49cc-b5f1-ff1d8addce42" containerName="pull" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.393910 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b524d641-ed92-49cc-b5f1-ff1d8addce42" containerName="pull" Oct 07 13:57:29 crc kubenswrapper[4959]: E1007 13:57:29.393917 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fca2a0e-545f-43b0-b4ec-e5f15babcd71" containerName="console" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.393923 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fca2a0e-545f-43b0-b4ec-e5f15babcd71" containerName="console" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.394017 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b524d641-ed92-49cc-b5f1-ff1d8addce42" containerName="extract" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.394024 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fca2a0e-545f-43b0-b4ec-e5f15babcd71" containerName="console" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.394440 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-9c45bf8cc-m9pg5" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.398578 4959 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.398757 4959 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.398899 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.400700 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.400860 4959 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-whkm6" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.428673 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dce9e35b-f4a5-41c7-a47f-f9de5ac4d966-webhook-cert\") pod \"metallb-operator-controller-manager-9c45bf8cc-m9pg5\" (UID: \"dce9e35b-f4a5-41c7-a47f-f9de5ac4d966\") " pod="metallb-system/metallb-operator-controller-manager-9c45bf8cc-m9pg5" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.428745 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrj22\" (UniqueName: \"kubernetes.io/projected/dce9e35b-f4a5-41c7-a47f-f9de5ac4d966-kube-api-access-qrj22\") pod \"metallb-operator-controller-manager-9c45bf8cc-m9pg5\" (UID: \"dce9e35b-f4a5-41c7-a47f-f9de5ac4d966\") " pod="metallb-system/metallb-operator-controller-manager-9c45bf8cc-m9pg5" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.428945 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dce9e35b-f4a5-41c7-a47f-f9de5ac4d966-apiservice-cert\") pod \"metallb-operator-controller-manager-9c45bf8cc-m9pg5\" (UID: \"dce9e35b-f4a5-41c7-a47f-f9de5ac4d966\") " pod="metallb-system/metallb-operator-controller-manager-9c45bf8cc-m9pg5" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.456519 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-9c45bf8cc-m9pg5"] Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.531031 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dce9e35b-f4a5-41c7-a47f-f9de5ac4d966-webhook-cert\") pod \"metallb-operator-controller-manager-9c45bf8cc-m9pg5\" (UID: \"dce9e35b-f4a5-41c7-a47f-f9de5ac4d966\") " pod="metallb-system/metallb-operator-controller-manager-9c45bf8cc-m9pg5" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.531155 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrj22\" (UniqueName: \"kubernetes.io/projected/dce9e35b-f4a5-41c7-a47f-f9de5ac4d966-kube-api-access-qrj22\") pod \"metallb-operator-controller-manager-9c45bf8cc-m9pg5\" (UID: \"dce9e35b-f4a5-41c7-a47f-f9de5ac4d966\") " pod="metallb-system/metallb-operator-controller-manager-9c45bf8cc-m9pg5" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.531198 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dce9e35b-f4a5-41c7-a47f-f9de5ac4d966-apiservice-cert\") pod \"metallb-operator-controller-manager-9c45bf8cc-m9pg5\" (UID: \"dce9e35b-f4a5-41c7-a47f-f9de5ac4d966\") " pod="metallb-system/metallb-operator-controller-manager-9c45bf8cc-m9pg5" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.537530 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dce9e35b-f4a5-41c7-a47f-f9de5ac4d966-apiservice-cert\") pod \"metallb-operator-controller-manager-9c45bf8cc-m9pg5\" (UID: \"dce9e35b-f4a5-41c7-a47f-f9de5ac4d966\") " pod="metallb-system/metallb-operator-controller-manager-9c45bf8cc-m9pg5" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.550032 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dce9e35b-f4a5-41c7-a47f-f9de5ac4d966-webhook-cert\") pod \"metallb-operator-controller-manager-9c45bf8cc-m9pg5\" (UID: \"dce9e35b-f4a5-41c7-a47f-f9de5ac4d966\") " pod="metallb-system/metallb-operator-controller-manager-9c45bf8cc-m9pg5" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.555874 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrj22\" (UniqueName: \"kubernetes.io/projected/dce9e35b-f4a5-41c7-a47f-f9de5ac4d966-kube-api-access-qrj22\") pod \"metallb-operator-controller-manager-9c45bf8cc-m9pg5\" (UID: \"dce9e35b-f4a5-41c7-a47f-f9de5ac4d966\") " pod="metallb-system/metallb-operator-controller-manager-9c45bf8cc-m9pg5" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.693979 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-7d969dc76f-ss6c5"] Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.694656 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7d969dc76f-ss6c5" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.697348 4959 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.697457 4959 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-98vdt" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.698399 4959 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.707389 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7d969dc76f-ss6c5"] Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.713425 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-9c45bf8cc-m9pg5" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.733543 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6a48c903-5280-4552-813c-47799ec95dfa-apiservice-cert\") pod \"metallb-operator-webhook-server-7d969dc76f-ss6c5\" (UID: \"6a48c903-5280-4552-813c-47799ec95dfa\") " pod="metallb-system/metallb-operator-webhook-server-7d969dc76f-ss6c5" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.733638 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcw5g\" (UniqueName: \"kubernetes.io/projected/6a48c903-5280-4552-813c-47799ec95dfa-kube-api-access-wcw5g\") pod \"metallb-operator-webhook-server-7d969dc76f-ss6c5\" (UID: \"6a48c903-5280-4552-813c-47799ec95dfa\") " pod="metallb-system/metallb-operator-webhook-server-7d969dc76f-ss6c5" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.733666 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6a48c903-5280-4552-813c-47799ec95dfa-webhook-cert\") pod \"metallb-operator-webhook-server-7d969dc76f-ss6c5\" (UID: \"6a48c903-5280-4552-813c-47799ec95dfa\") " pod="metallb-system/metallb-operator-webhook-server-7d969dc76f-ss6c5" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.835070 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcw5g\" (UniqueName: \"kubernetes.io/projected/6a48c903-5280-4552-813c-47799ec95dfa-kube-api-access-wcw5g\") pod \"metallb-operator-webhook-server-7d969dc76f-ss6c5\" (UID: \"6a48c903-5280-4552-813c-47799ec95dfa\") " pod="metallb-system/metallb-operator-webhook-server-7d969dc76f-ss6c5" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.835136 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6a48c903-5280-4552-813c-47799ec95dfa-webhook-cert\") pod \"metallb-operator-webhook-server-7d969dc76f-ss6c5\" (UID: \"6a48c903-5280-4552-813c-47799ec95dfa\") " pod="metallb-system/metallb-operator-webhook-server-7d969dc76f-ss6c5" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.835162 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6a48c903-5280-4552-813c-47799ec95dfa-apiservice-cert\") pod \"metallb-operator-webhook-server-7d969dc76f-ss6c5\" (UID: \"6a48c903-5280-4552-813c-47799ec95dfa\") " pod="metallb-system/metallb-operator-webhook-server-7d969dc76f-ss6c5" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.838726 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6a48c903-5280-4552-813c-47799ec95dfa-apiservice-cert\") pod \"metallb-operator-webhook-server-7d969dc76f-ss6c5\" (UID: \"6a48c903-5280-4552-813c-47799ec95dfa\") " pod="metallb-system/metallb-operator-webhook-server-7d969dc76f-ss6c5" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.841624 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6a48c903-5280-4552-813c-47799ec95dfa-webhook-cert\") pod \"metallb-operator-webhook-server-7d969dc76f-ss6c5\" (UID: \"6a48c903-5280-4552-813c-47799ec95dfa\") " pod="metallb-system/metallb-operator-webhook-server-7d969dc76f-ss6c5" Oct 07 13:57:29 crc kubenswrapper[4959]: I1007 13:57:29.853183 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcw5g\" (UniqueName: \"kubernetes.io/projected/6a48c903-5280-4552-813c-47799ec95dfa-kube-api-access-wcw5g\") pod \"metallb-operator-webhook-server-7d969dc76f-ss6c5\" (UID: \"6a48c903-5280-4552-813c-47799ec95dfa\") " pod="metallb-system/metallb-operator-webhook-server-7d969dc76f-ss6c5" Oct 07 13:57:30 crc kubenswrapper[4959]: I1007 13:57:30.009781 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7d969dc76f-ss6c5" Oct 07 13:57:30 crc kubenswrapper[4959]: I1007 13:57:30.026805 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-9c45bf8cc-m9pg5"] Oct 07 13:57:30 crc kubenswrapper[4959]: W1007 13:57:30.038398 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddce9e35b_f4a5_41c7_a47f_f9de5ac4d966.slice/crio-edd1bd165d171de0848d60a72121dc5dbd2eb5c200bdba0d1d6fbe595cd613c7 WatchSource:0}: Error finding container edd1bd165d171de0848d60a72121dc5dbd2eb5c200bdba0d1d6fbe595cd613c7: Status 404 returned error can't find the container with id edd1bd165d171de0848d60a72121dc5dbd2eb5c200bdba0d1d6fbe595cd613c7 Oct 07 13:57:30 crc kubenswrapper[4959]: I1007 13:57:30.232314 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7d969dc76f-ss6c5"] Oct 07 13:57:30 crc kubenswrapper[4959]: W1007 13:57:30.240828 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a48c903_5280_4552_813c_47799ec95dfa.slice/crio-485b2d6f64a805dfe62b3f801e2113664365e4fb06d56db3c5b02fca5f497eb3 WatchSource:0}: Error finding container 485b2d6f64a805dfe62b3f801e2113664365e4fb06d56db3c5b02fca5f497eb3: Status 404 returned error can't find the container with id 485b2d6f64a805dfe62b3f801e2113664365e4fb06d56db3c5b02fca5f497eb3 Oct 07 13:57:30 crc kubenswrapper[4959]: I1007 13:57:30.378024 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-9c45bf8cc-m9pg5" event={"ID":"dce9e35b-f4a5-41c7-a47f-f9de5ac4d966","Type":"ContainerStarted","Data":"edd1bd165d171de0848d60a72121dc5dbd2eb5c200bdba0d1d6fbe595cd613c7"} Oct 07 13:57:30 crc kubenswrapper[4959]: I1007 13:57:30.379466 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7d969dc76f-ss6c5" event={"ID":"6a48c903-5280-4552-813c-47799ec95dfa","Type":"ContainerStarted","Data":"485b2d6f64a805dfe62b3f801e2113664365e4fb06d56db3c5b02fca5f497eb3"} Oct 07 13:57:35 crc kubenswrapper[4959]: I1007 13:57:35.412839 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-9c45bf8cc-m9pg5" event={"ID":"dce9e35b-f4a5-41c7-a47f-f9de5ac4d966","Type":"ContainerStarted","Data":"56633c5ec46db7311897bb84e824a0a7d97f7487d1ce30d5550fb6aac5ff0aac"} Oct 07 13:57:35 crc kubenswrapper[4959]: I1007 13:57:35.413513 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-9c45bf8cc-m9pg5" Oct 07 13:57:35 crc kubenswrapper[4959]: I1007 13:57:35.415599 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7d969dc76f-ss6c5" event={"ID":"6a48c903-5280-4552-813c-47799ec95dfa","Type":"ContainerStarted","Data":"974fab27a43a48e8aa1f1320683b223e7d6bd9265ead67e5cbf338d185a2c10d"} Oct 07 13:57:35 crc kubenswrapper[4959]: I1007 13:57:35.415808 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-7d969dc76f-ss6c5" Oct 07 13:57:35 crc kubenswrapper[4959]: I1007 13:57:35.432606 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-9c45bf8cc-m9pg5" podStartSLOduration=1.555455653 podStartE2EDuration="6.432587514s" podCreationTimestamp="2025-10-07 13:57:29 +0000 UTC" firstStartedPulling="2025-10-07 13:57:30.042622333 +0000 UTC m=+732.126026658" lastFinishedPulling="2025-10-07 13:57:34.919754184 +0000 UTC m=+737.003158519" observedRunningTime="2025-10-07 13:57:35.429985508 +0000 UTC m=+737.513389833" watchObservedRunningTime="2025-10-07 13:57:35.432587514 +0000 UTC m=+737.515991839" Oct 07 13:57:35 crc kubenswrapper[4959]: I1007 13:57:35.462215 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-7d969dc76f-ss6c5" podStartSLOduration=1.775319653 podStartE2EDuration="6.462175752s" podCreationTimestamp="2025-10-07 13:57:29 +0000 UTC" firstStartedPulling="2025-10-07 13:57:30.247225687 +0000 UTC m=+732.330630012" lastFinishedPulling="2025-10-07 13:57:34.934081776 +0000 UTC m=+737.017486111" observedRunningTime="2025-10-07 13:57:35.456767655 +0000 UTC m=+737.540172010" watchObservedRunningTime="2025-10-07 13:57:35.462175752 +0000 UTC m=+737.545580117" Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.335415 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9drmk"] Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.336352 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-9drmk" podUID="c759802a-33e9-49e6-b764-7504e867562f" containerName="controller-manager" containerID="cri-o://0053e3376bc6e7a8a2a6accdb09ba29008fd73067a965a6b29ce48e43a69f3df" gracePeriod=30 Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.350654 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87"] Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.350856 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87" podUID="0ff428ab-6c11-4571-9717-9ac0bfe462d4" containerName="route-controller-manager" containerID="cri-o://91dc56d48a2de7414ced1f338d667518a1ffdd206b127332624a0f63607898dc" gracePeriod=30 Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.474418 4959 generic.go:334] "Generic (PLEG): container finished" podID="c759802a-33e9-49e6-b764-7504e867562f" containerID="0053e3376bc6e7a8a2a6accdb09ba29008fd73067a965a6b29ce48e43a69f3df" exitCode=0 Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.474605 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9drmk" event={"ID":"c759802a-33e9-49e6-b764-7504e867562f","Type":"ContainerDied","Data":"0053e3376bc6e7a8a2a6accdb09ba29008fd73067a965a6b29ce48e43a69f3df"} Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.787309 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9drmk" Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.845402 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87" Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.861084 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c759802a-33e9-49e6-b764-7504e867562f-proxy-ca-bundles\") pod \"c759802a-33e9-49e6-b764-7504e867562f\" (UID: \"c759802a-33e9-49e6-b764-7504e867562f\") " Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.861158 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c759802a-33e9-49e6-b764-7504e867562f-serving-cert\") pod \"c759802a-33e9-49e6-b764-7504e867562f\" (UID: \"c759802a-33e9-49e6-b764-7504e867562f\") " Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.861205 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c759802a-33e9-49e6-b764-7504e867562f-client-ca\") pod \"c759802a-33e9-49e6-b764-7504e867562f\" (UID: \"c759802a-33e9-49e6-b764-7504e867562f\") " Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.861246 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jxg6\" (UniqueName: \"kubernetes.io/projected/c759802a-33e9-49e6-b764-7504e867562f-kube-api-access-4jxg6\") pod \"c759802a-33e9-49e6-b764-7504e867562f\" (UID: \"c759802a-33e9-49e6-b764-7504e867562f\") " Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.861271 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c759802a-33e9-49e6-b764-7504e867562f-config\") pod \"c759802a-33e9-49e6-b764-7504e867562f\" (UID: \"c759802a-33e9-49e6-b764-7504e867562f\") " Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.862256 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c759802a-33e9-49e6-b764-7504e867562f-config" (OuterVolumeSpecName: "config") pod "c759802a-33e9-49e6-b764-7504e867562f" (UID: "c759802a-33e9-49e6-b764-7504e867562f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.862708 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c759802a-33e9-49e6-b764-7504e867562f-client-ca" (OuterVolumeSpecName: "client-ca") pod "c759802a-33e9-49e6-b764-7504e867562f" (UID: "c759802a-33e9-49e6-b764-7504e867562f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.863113 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c759802a-33e9-49e6-b764-7504e867562f-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "c759802a-33e9-49e6-b764-7504e867562f" (UID: "c759802a-33e9-49e6-b764-7504e867562f"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.866839 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c759802a-33e9-49e6-b764-7504e867562f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c759802a-33e9-49e6-b764-7504e867562f" (UID: "c759802a-33e9-49e6-b764-7504e867562f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.867373 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c759802a-33e9-49e6-b764-7504e867562f-kube-api-access-4jxg6" (OuterVolumeSpecName: "kube-api-access-4jxg6") pod "c759802a-33e9-49e6-b764-7504e867562f" (UID: "c759802a-33e9-49e6-b764-7504e867562f"). InnerVolumeSpecName "kube-api-access-4jxg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.962746 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ff428ab-6c11-4571-9717-9ac0bfe462d4-serving-cert\") pod \"0ff428ab-6c11-4571-9717-9ac0bfe462d4\" (UID: \"0ff428ab-6c11-4571-9717-9ac0bfe462d4\") " Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.962916 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvcdd\" (UniqueName: \"kubernetes.io/projected/0ff428ab-6c11-4571-9717-9ac0bfe462d4-kube-api-access-xvcdd\") pod \"0ff428ab-6c11-4571-9717-9ac0bfe462d4\" (UID: \"0ff428ab-6c11-4571-9717-9ac0bfe462d4\") " Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.962948 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0ff428ab-6c11-4571-9717-9ac0bfe462d4-client-ca\") pod \"0ff428ab-6c11-4571-9717-9ac0bfe462d4\" (UID: \"0ff428ab-6c11-4571-9717-9ac0bfe462d4\") " Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.962969 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ff428ab-6c11-4571-9717-9ac0bfe462d4-config\") pod \"0ff428ab-6c11-4571-9717-9ac0bfe462d4\" (UID: \"0ff428ab-6c11-4571-9717-9ac0bfe462d4\") " Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.963183 4959 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c759802a-33e9-49e6-b764-7504e867562f-client-ca\") on node \"crc\" DevicePath \"\"" Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.963199 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jxg6\" (UniqueName: \"kubernetes.io/projected/c759802a-33e9-49e6-b764-7504e867562f-kube-api-access-4jxg6\") on node \"crc\" DevicePath \"\"" Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.963209 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c759802a-33e9-49e6-b764-7504e867562f-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.963218 4959 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c759802a-33e9-49e6-b764-7504e867562f-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.963233 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c759802a-33e9-49e6-b764-7504e867562f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.963683 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ff428ab-6c11-4571-9717-9ac0bfe462d4-client-ca" (OuterVolumeSpecName: "client-ca") pod "0ff428ab-6c11-4571-9717-9ac0bfe462d4" (UID: "0ff428ab-6c11-4571-9717-9ac0bfe462d4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.963717 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ff428ab-6c11-4571-9717-9ac0bfe462d4-config" (OuterVolumeSpecName: "config") pod "0ff428ab-6c11-4571-9717-9ac0bfe462d4" (UID: "0ff428ab-6c11-4571-9717-9ac0bfe462d4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.967380 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ff428ab-6c11-4571-9717-9ac0bfe462d4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0ff428ab-6c11-4571-9717-9ac0bfe462d4" (UID: "0ff428ab-6c11-4571-9717-9ac0bfe462d4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 13:57:44 crc kubenswrapper[4959]: I1007 13:57:44.972348 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ff428ab-6c11-4571-9717-9ac0bfe462d4-kube-api-access-xvcdd" (OuterVolumeSpecName: "kube-api-access-xvcdd") pod "0ff428ab-6c11-4571-9717-9ac0bfe462d4" (UID: "0ff428ab-6c11-4571-9717-9ac0bfe462d4"). InnerVolumeSpecName "kube-api-access-xvcdd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:57:45 crc kubenswrapper[4959]: I1007 13:57:45.064725 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvcdd\" (UniqueName: \"kubernetes.io/projected/0ff428ab-6c11-4571-9717-9ac0bfe462d4-kube-api-access-xvcdd\") on node \"crc\" DevicePath \"\"" Oct 07 13:57:45 crc kubenswrapper[4959]: I1007 13:57:45.065044 4959 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0ff428ab-6c11-4571-9717-9ac0bfe462d4-client-ca\") on node \"crc\" DevicePath \"\"" Oct 07 13:57:45 crc kubenswrapper[4959]: I1007 13:57:45.065056 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ff428ab-6c11-4571-9717-9ac0bfe462d4-config\") on node \"crc\" DevicePath \"\"" Oct 07 13:57:45 crc kubenswrapper[4959]: I1007 13:57:45.065066 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ff428ab-6c11-4571-9717-9ac0bfe462d4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 07 13:57:45 crc kubenswrapper[4959]: I1007 13:57:45.483883 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9drmk" event={"ID":"c759802a-33e9-49e6-b764-7504e867562f","Type":"ContainerDied","Data":"3f94177413fca52e34dc650f9cab31a0b80c13a88f4fbc3dd47b6024fec3f45e"} Oct 07 13:57:45 crc kubenswrapper[4959]: I1007 13:57:45.483938 4959 scope.go:117] "RemoveContainer" containerID="0053e3376bc6e7a8a2a6accdb09ba29008fd73067a965a6b29ce48e43a69f3df" Oct 07 13:57:45 crc kubenswrapper[4959]: I1007 13:57:45.483941 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9drmk" Oct 07 13:57:45 crc kubenswrapper[4959]: I1007 13:57:45.485455 4959 generic.go:334] "Generic (PLEG): container finished" podID="0ff428ab-6c11-4571-9717-9ac0bfe462d4" containerID="91dc56d48a2de7414ced1f338d667518a1ffdd206b127332624a0f63607898dc" exitCode=0 Oct 07 13:57:45 crc kubenswrapper[4959]: I1007 13:57:45.485498 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87" event={"ID":"0ff428ab-6c11-4571-9717-9ac0bfe462d4","Type":"ContainerDied","Data":"91dc56d48a2de7414ced1f338d667518a1ffdd206b127332624a0f63607898dc"} Oct 07 13:57:45 crc kubenswrapper[4959]: I1007 13:57:45.485533 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87" event={"ID":"0ff428ab-6c11-4571-9717-9ac0bfe462d4","Type":"ContainerDied","Data":"0db27f84e0fc2074ebc40d626583ccda52be1b4d22894a57733d49ebe30d3816"} Oct 07 13:57:45 crc kubenswrapper[4959]: I1007 13:57:45.485518 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87" Oct 07 13:57:45 crc kubenswrapper[4959]: I1007 13:57:45.511925 4959 scope.go:117] "RemoveContainer" containerID="91dc56d48a2de7414ced1f338d667518a1ffdd206b127332624a0f63607898dc" Oct 07 13:57:45 crc kubenswrapper[4959]: I1007 13:57:45.525200 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9drmk"] Oct 07 13:57:45 crc kubenswrapper[4959]: I1007 13:57:45.525624 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9drmk"] Oct 07 13:57:45 crc kubenswrapper[4959]: I1007 13:57:45.527559 4959 scope.go:117] "RemoveContainer" containerID="91dc56d48a2de7414ced1f338d667518a1ffdd206b127332624a0f63607898dc" Oct 07 13:57:45 crc kubenswrapper[4959]: E1007 13:57:45.528046 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91dc56d48a2de7414ced1f338d667518a1ffdd206b127332624a0f63607898dc\": container with ID starting with 91dc56d48a2de7414ced1f338d667518a1ffdd206b127332624a0f63607898dc not found: ID does not exist" containerID="91dc56d48a2de7414ced1f338d667518a1ffdd206b127332624a0f63607898dc" Oct 07 13:57:45 crc kubenswrapper[4959]: I1007 13:57:45.528085 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91dc56d48a2de7414ced1f338d667518a1ffdd206b127332624a0f63607898dc"} err="failed to get container status \"91dc56d48a2de7414ced1f338d667518a1ffdd206b127332624a0f63607898dc\": rpc error: code = NotFound desc = could not find container \"91dc56d48a2de7414ced1f338d667518a1ffdd206b127332624a0f63607898dc\": container with ID starting with 91dc56d48a2de7414ced1f338d667518a1ffdd206b127332624a0f63607898dc not found: ID does not exist" Oct 07 13:57:45 crc kubenswrapper[4959]: I1007 13:57:45.535916 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87"] Oct 07 13:57:45 crc kubenswrapper[4959]: I1007 13:57:45.538893 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wfh87"] Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.320940 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-55f65b4f96-j28sc"] Oct 07 13:57:46 crc kubenswrapper[4959]: E1007 13:57:46.321191 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ff428ab-6c11-4571-9717-9ac0bfe462d4" containerName="route-controller-manager" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.321206 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ff428ab-6c11-4571-9717-9ac0bfe462d4" containerName="route-controller-manager" Oct 07 13:57:46 crc kubenswrapper[4959]: E1007 13:57:46.321228 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c759802a-33e9-49e6-b764-7504e867562f" containerName="controller-manager" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.321235 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c759802a-33e9-49e6-b764-7504e867562f" containerName="controller-manager" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.321353 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="c759802a-33e9-49e6-b764-7504e867562f" containerName="controller-manager" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.321375 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ff428ab-6c11-4571-9717-9ac0bfe462d4" containerName="route-controller-manager" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.321868 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-55f65b4f96-j28sc" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.327189 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-779cd7b465-96ws6"] Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.327994 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-779cd7b465-96ws6" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.330768 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.330845 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.330796 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.331158 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.331227 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.331370 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.331627 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.331630 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.331710 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.331823 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.332345 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.332514 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.334447 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-55f65b4f96-j28sc"] Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.338319 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.348682 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-779cd7b465-96ws6"] Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.393919 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f4f7c1ae-d966-41ef-b43d-b387174767cf-client-ca\") pod \"route-controller-manager-779cd7b465-96ws6\" (UID: \"f4f7c1ae-d966-41ef-b43d-b387174767cf\") " pod="openshift-route-controller-manager/route-controller-manager-779cd7b465-96ws6" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.393979 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4f7c1ae-d966-41ef-b43d-b387174767cf-serving-cert\") pod \"route-controller-manager-779cd7b465-96ws6\" (UID: \"f4f7c1ae-d966-41ef-b43d-b387174767cf\") " pod="openshift-route-controller-manager/route-controller-manager-779cd7b465-96ws6" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.394011 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9fj5\" (UniqueName: \"kubernetes.io/projected/f4f7c1ae-d966-41ef-b43d-b387174767cf-kube-api-access-b9fj5\") pod \"route-controller-manager-779cd7b465-96ws6\" (UID: \"f4f7c1ae-d966-41ef-b43d-b387174767cf\") " pod="openshift-route-controller-manager/route-controller-manager-779cd7b465-96ws6" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.394042 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b4646ca2-f125-429b-a94b-76eac50769a6-client-ca\") pod \"controller-manager-55f65b4f96-j28sc\" (UID: \"b4646ca2-f125-429b-a94b-76eac50769a6\") " pod="openshift-controller-manager/controller-manager-55f65b4f96-j28sc" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.394171 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4646ca2-f125-429b-a94b-76eac50769a6-config\") pod \"controller-manager-55f65b4f96-j28sc\" (UID: \"b4646ca2-f125-429b-a94b-76eac50769a6\") " pod="openshift-controller-manager/controller-manager-55f65b4f96-j28sc" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.394339 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvdr4\" (UniqueName: \"kubernetes.io/projected/b4646ca2-f125-429b-a94b-76eac50769a6-kube-api-access-wvdr4\") pod \"controller-manager-55f65b4f96-j28sc\" (UID: \"b4646ca2-f125-429b-a94b-76eac50769a6\") " pod="openshift-controller-manager/controller-manager-55f65b4f96-j28sc" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.394386 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b4646ca2-f125-429b-a94b-76eac50769a6-proxy-ca-bundles\") pod \"controller-manager-55f65b4f96-j28sc\" (UID: \"b4646ca2-f125-429b-a94b-76eac50769a6\") " pod="openshift-controller-manager/controller-manager-55f65b4f96-j28sc" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.394483 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4646ca2-f125-429b-a94b-76eac50769a6-serving-cert\") pod \"controller-manager-55f65b4f96-j28sc\" (UID: \"b4646ca2-f125-429b-a94b-76eac50769a6\") " pod="openshift-controller-manager/controller-manager-55f65b4f96-j28sc" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.394564 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4f7c1ae-d966-41ef-b43d-b387174767cf-config\") pod \"route-controller-manager-779cd7b465-96ws6\" (UID: \"f4f7c1ae-d966-41ef-b43d-b387174767cf\") " pod="openshift-route-controller-manager/route-controller-manager-779cd7b465-96ws6" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.496215 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4646ca2-f125-429b-a94b-76eac50769a6-serving-cert\") pod \"controller-manager-55f65b4f96-j28sc\" (UID: \"b4646ca2-f125-429b-a94b-76eac50769a6\") " pod="openshift-controller-manager/controller-manager-55f65b4f96-j28sc" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.496330 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4f7c1ae-d966-41ef-b43d-b387174767cf-config\") pod \"route-controller-manager-779cd7b465-96ws6\" (UID: \"f4f7c1ae-d966-41ef-b43d-b387174767cf\") " pod="openshift-route-controller-manager/route-controller-manager-779cd7b465-96ws6" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.496378 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f4f7c1ae-d966-41ef-b43d-b387174767cf-client-ca\") pod \"route-controller-manager-779cd7b465-96ws6\" (UID: \"f4f7c1ae-d966-41ef-b43d-b387174767cf\") " pod="openshift-route-controller-manager/route-controller-manager-779cd7b465-96ws6" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.496403 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4f7c1ae-d966-41ef-b43d-b387174767cf-serving-cert\") pod \"route-controller-manager-779cd7b465-96ws6\" (UID: \"f4f7c1ae-d966-41ef-b43d-b387174767cf\") " pod="openshift-route-controller-manager/route-controller-manager-779cd7b465-96ws6" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.497169 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9fj5\" (UniqueName: \"kubernetes.io/projected/f4f7c1ae-d966-41ef-b43d-b387174767cf-kube-api-access-b9fj5\") pod \"route-controller-manager-779cd7b465-96ws6\" (UID: \"f4f7c1ae-d966-41ef-b43d-b387174767cf\") " pod="openshift-route-controller-manager/route-controller-manager-779cd7b465-96ws6" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.497239 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b4646ca2-f125-429b-a94b-76eac50769a6-client-ca\") pod \"controller-manager-55f65b4f96-j28sc\" (UID: \"b4646ca2-f125-429b-a94b-76eac50769a6\") " pod="openshift-controller-manager/controller-manager-55f65b4f96-j28sc" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.497337 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4646ca2-f125-429b-a94b-76eac50769a6-config\") pod \"controller-manager-55f65b4f96-j28sc\" (UID: \"b4646ca2-f125-429b-a94b-76eac50769a6\") " pod="openshift-controller-manager/controller-manager-55f65b4f96-j28sc" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.497405 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvdr4\" (UniqueName: \"kubernetes.io/projected/b4646ca2-f125-429b-a94b-76eac50769a6-kube-api-access-wvdr4\") pod \"controller-manager-55f65b4f96-j28sc\" (UID: \"b4646ca2-f125-429b-a94b-76eac50769a6\") " pod="openshift-controller-manager/controller-manager-55f65b4f96-j28sc" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.497336 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f4f7c1ae-d966-41ef-b43d-b387174767cf-client-ca\") pod \"route-controller-manager-779cd7b465-96ws6\" (UID: \"f4f7c1ae-d966-41ef-b43d-b387174767cf\") " pod="openshift-route-controller-manager/route-controller-manager-779cd7b465-96ws6" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.497454 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4f7c1ae-d966-41ef-b43d-b387174767cf-config\") pod \"route-controller-manager-779cd7b465-96ws6\" (UID: \"f4f7c1ae-d966-41ef-b43d-b387174767cf\") " pod="openshift-route-controller-manager/route-controller-manager-779cd7b465-96ws6" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.497439 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b4646ca2-f125-429b-a94b-76eac50769a6-proxy-ca-bundles\") pod \"controller-manager-55f65b4f96-j28sc\" (UID: \"b4646ca2-f125-429b-a94b-76eac50769a6\") " pod="openshift-controller-manager/controller-manager-55f65b4f96-j28sc" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.498176 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b4646ca2-f125-429b-a94b-76eac50769a6-client-ca\") pod \"controller-manager-55f65b4f96-j28sc\" (UID: \"b4646ca2-f125-429b-a94b-76eac50769a6\") " pod="openshift-controller-manager/controller-manager-55f65b4f96-j28sc" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.501902 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b4646ca2-f125-429b-a94b-76eac50769a6-proxy-ca-bundles\") pod \"controller-manager-55f65b4f96-j28sc\" (UID: \"b4646ca2-f125-429b-a94b-76eac50769a6\") " pod="openshift-controller-manager/controller-manager-55f65b4f96-j28sc" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.501917 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4f7c1ae-d966-41ef-b43d-b387174767cf-serving-cert\") pod \"route-controller-manager-779cd7b465-96ws6\" (UID: \"f4f7c1ae-d966-41ef-b43d-b387174767cf\") " pod="openshift-route-controller-manager/route-controller-manager-779cd7b465-96ws6" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.501991 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4646ca2-f125-429b-a94b-76eac50769a6-serving-cert\") pod \"controller-manager-55f65b4f96-j28sc\" (UID: \"b4646ca2-f125-429b-a94b-76eac50769a6\") " pod="openshift-controller-manager/controller-manager-55f65b4f96-j28sc" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.502345 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4646ca2-f125-429b-a94b-76eac50769a6-config\") pod \"controller-manager-55f65b4f96-j28sc\" (UID: \"b4646ca2-f125-429b-a94b-76eac50769a6\") " pod="openshift-controller-manager/controller-manager-55f65b4f96-j28sc" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.515727 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvdr4\" (UniqueName: \"kubernetes.io/projected/b4646ca2-f125-429b-a94b-76eac50769a6-kube-api-access-wvdr4\") pod \"controller-manager-55f65b4f96-j28sc\" (UID: \"b4646ca2-f125-429b-a94b-76eac50769a6\") " pod="openshift-controller-manager/controller-manager-55f65b4f96-j28sc" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.519252 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9fj5\" (UniqueName: \"kubernetes.io/projected/f4f7c1ae-d966-41ef-b43d-b387174767cf-kube-api-access-b9fj5\") pod \"route-controller-manager-779cd7b465-96ws6\" (UID: \"f4f7c1ae-d966-41ef-b43d-b387174767cf\") " pod="openshift-route-controller-manager/route-controller-manager-779cd7b465-96ws6" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.666944 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-55f65b4f96-j28sc" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.669567 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ff428ab-6c11-4571-9717-9ac0bfe462d4" path="/var/lib/kubelet/pods/0ff428ab-6c11-4571-9717-9ac0bfe462d4/volumes" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.670827 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c759802a-33e9-49e6-b764-7504e867562f" path="/var/lib/kubelet/pods/c759802a-33e9-49e6-b764-7504e867562f/volumes" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.678339 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-779cd7b465-96ws6" Oct 07 13:57:46 crc kubenswrapper[4959]: I1007 13:57:46.915745 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-55f65b4f96-j28sc"] Oct 07 13:57:47 crc kubenswrapper[4959]: I1007 13:57:47.193518 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-779cd7b465-96ws6"] Oct 07 13:57:47 crc kubenswrapper[4959]: W1007 13:57:47.214817 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4f7c1ae_d966_41ef_b43d_b387174767cf.slice/crio-92390ff0c497e88051c546360d2396084d6eee18c622ce41f7f46380fbd13b55 WatchSource:0}: Error finding container 92390ff0c497e88051c546360d2396084d6eee18c622ce41f7f46380fbd13b55: Status 404 returned error can't find the container with id 92390ff0c497e88051c546360d2396084d6eee18c622ce41f7f46380fbd13b55 Oct 07 13:57:47 crc kubenswrapper[4959]: I1007 13:57:47.498414 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-55f65b4f96-j28sc" event={"ID":"b4646ca2-f125-429b-a94b-76eac50769a6","Type":"ContainerStarted","Data":"c0d32cb6171e096532eb6cbf1313a4e9d80e422b85173b611ab955ce61720c00"} Oct 07 13:57:47 crc kubenswrapper[4959]: I1007 13:57:47.499683 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-55f65b4f96-j28sc" event={"ID":"b4646ca2-f125-429b-a94b-76eac50769a6","Type":"ContainerStarted","Data":"8aae08744dddc67431f2a0ff23593ab9ba7a3c3b55e78ab4b358e4e602429f17"} Oct 07 13:57:47 crc kubenswrapper[4959]: I1007 13:57:47.499945 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-55f65b4f96-j28sc" Oct 07 13:57:47 crc kubenswrapper[4959]: I1007 13:57:47.500581 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-779cd7b465-96ws6" event={"ID":"f4f7c1ae-d966-41ef-b43d-b387174767cf","Type":"ContainerStarted","Data":"11cef174fbaeeb8eee49b87b5835f730b3477928683f3ac0ad96fe36bbe0fefd"} Oct 07 13:57:47 crc kubenswrapper[4959]: I1007 13:57:47.500695 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-779cd7b465-96ws6" event={"ID":"f4f7c1ae-d966-41ef-b43d-b387174767cf","Type":"ContainerStarted","Data":"92390ff0c497e88051c546360d2396084d6eee18c622ce41f7f46380fbd13b55"} Oct 07 13:57:47 crc kubenswrapper[4959]: I1007 13:57:47.501516 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-779cd7b465-96ws6" Oct 07 13:57:47 crc kubenswrapper[4959]: I1007 13:57:47.503256 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-55f65b4f96-j28sc" Oct 07 13:57:47 crc kubenswrapper[4959]: I1007 13:57:47.522574 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-55f65b4f96-j28sc" podStartSLOduration=3.522559118 podStartE2EDuration="3.522559118s" podCreationTimestamp="2025-10-07 13:57:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:57:47.521388619 +0000 UTC m=+749.604792954" watchObservedRunningTime="2025-10-07 13:57:47.522559118 +0000 UTC m=+749.605963443" Oct 07 13:57:47 crc kubenswrapper[4959]: I1007 13:57:47.536017 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-779cd7b465-96ws6" podStartSLOduration=3.5360038879999998 podStartE2EDuration="3.536003888s" podCreationTimestamp="2025-10-07 13:57:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:57:47.535334331 +0000 UTC m=+749.618738656" watchObservedRunningTime="2025-10-07 13:57:47.536003888 +0000 UTC m=+749.619408213" Oct 07 13:57:47 crc kubenswrapper[4959]: I1007 13:57:47.801662 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-779cd7b465-96ws6" Oct 07 13:57:50 crc kubenswrapper[4959]: I1007 13:57:50.016141 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-7d969dc76f-ss6c5" Oct 07 13:57:50 crc kubenswrapper[4959]: I1007 13:57:50.629929 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 13:57:50 crc kubenswrapper[4959]: I1007 13:57:50.629998 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 13:57:53 crc kubenswrapper[4959]: I1007 13:57:53.797835 4959 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 07 13:58:09 crc kubenswrapper[4959]: I1007 13:58:09.716355 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-9c45bf8cc-m9pg5" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.499638 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-sv4vw"] Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.501962 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-sv4vw" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.504802 4959 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.505026 4959 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-czsxs" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.505206 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.514351 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-wcc2g"] Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.515239 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wcc2g" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.516753 4959 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.528160 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-wcc2g"] Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.603997 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-bb2tj"] Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.604234 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6dd5ad64-5f6d-4e92-9c1b-3f0837cb134c-cert\") pod \"frr-k8s-webhook-server-64bf5d555-wcc2g\" (UID: \"6dd5ad64-5f6d-4e92-9c1b-3f0837cb134c\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wcc2g" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.604299 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zl8q\" (UniqueName: \"kubernetes.io/projected/5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f-kube-api-access-8zl8q\") pod \"frr-k8s-sv4vw\" (UID: \"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f\") " pod="metallb-system/frr-k8s-sv4vw" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.604323 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwddd\" (UniqueName: \"kubernetes.io/projected/6dd5ad64-5f6d-4e92-9c1b-3f0837cb134c-kube-api-access-nwddd\") pod \"frr-k8s-webhook-server-64bf5d555-wcc2g\" (UID: \"6dd5ad64-5f6d-4e92-9c1b-3f0837cb134c\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wcc2g" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.604339 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f-frr-sockets\") pod \"frr-k8s-sv4vw\" (UID: \"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f\") " pod="metallb-system/frr-k8s-sv4vw" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.604354 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f-metrics-certs\") pod \"frr-k8s-sv4vw\" (UID: \"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f\") " pod="metallb-system/frr-k8s-sv4vw" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.604372 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f-frr-startup\") pod \"frr-k8s-sv4vw\" (UID: \"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f\") " pod="metallb-system/frr-k8s-sv4vw" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.604397 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f-reloader\") pod \"frr-k8s-sv4vw\" (UID: \"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f\") " pod="metallb-system/frr-k8s-sv4vw" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.604413 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f-metrics\") pod \"frr-k8s-sv4vw\" (UID: \"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f\") " pod="metallb-system/frr-k8s-sv4vw" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.604444 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f-frr-conf\") pod \"frr-k8s-sv4vw\" (UID: \"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f\") " pod="metallb-system/frr-k8s-sv4vw" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.605236 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-bb2tj" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.613789 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-nbdf7"] Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.614688 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-nbdf7" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.615239 4959 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.615303 4959 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.615558 4959 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-g55gr" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.616172 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.616450 4959 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.628171 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-nbdf7"] Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.705245 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f-metrics\") pod \"frr-k8s-sv4vw\" (UID: \"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f\") " pod="metallb-system/frr-k8s-sv4vw" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.705308 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f-frr-conf\") pod \"frr-k8s-sv4vw\" (UID: \"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f\") " pod="metallb-system/frr-k8s-sv4vw" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.705339 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq7wl\" (UniqueName: \"kubernetes.io/projected/c9ac1753-f8ad-4f22-a955-38663b8121f0-kube-api-access-mq7wl\") pod \"speaker-bb2tj\" (UID: \"c9ac1753-f8ad-4f22-a955-38663b8121f0\") " pod="metallb-system/speaker-bb2tj" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.705375 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6dd5ad64-5f6d-4e92-9c1b-3f0837cb134c-cert\") pod \"frr-k8s-webhook-server-64bf5d555-wcc2g\" (UID: \"6dd5ad64-5f6d-4e92-9c1b-3f0837cb134c\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wcc2g" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.705409 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e2e00fa0-bc56-4a3f-9796-bbc1ec5dc47d-cert\") pod \"controller-68d546b9d8-nbdf7\" (UID: \"e2e00fa0-bc56-4a3f-9796-bbc1ec5dc47d\") " pod="metallb-system/controller-68d546b9d8-nbdf7" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.705434 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49hvz\" (UniqueName: \"kubernetes.io/projected/e2e00fa0-bc56-4a3f-9796-bbc1ec5dc47d-kube-api-access-49hvz\") pod \"controller-68d546b9d8-nbdf7\" (UID: \"e2e00fa0-bc56-4a3f-9796-bbc1ec5dc47d\") " pod="metallb-system/controller-68d546b9d8-nbdf7" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.705459 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e2e00fa0-bc56-4a3f-9796-bbc1ec5dc47d-metrics-certs\") pod \"controller-68d546b9d8-nbdf7\" (UID: \"e2e00fa0-bc56-4a3f-9796-bbc1ec5dc47d\") " pod="metallb-system/controller-68d546b9d8-nbdf7" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.705497 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c9ac1753-f8ad-4f22-a955-38663b8121f0-metrics-certs\") pod \"speaker-bb2tj\" (UID: \"c9ac1753-f8ad-4f22-a955-38663b8121f0\") " pod="metallb-system/speaker-bb2tj" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.705632 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zl8q\" (UniqueName: \"kubernetes.io/projected/5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f-kube-api-access-8zl8q\") pod \"frr-k8s-sv4vw\" (UID: \"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f\") " pod="metallb-system/frr-k8s-sv4vw" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.705660 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f-frr-sockets\") pod \"frr-k8s-sv4vw\" (UID: \"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f\") " pod="metallb-system/frr-k8s-sv4vw" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.705685 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwddd\" (UniqueName: \"kubernetes.io/projected/6dd5ad64-5f6d-4e92-9c1b-3f0837cb134c-kube-api-access-nwddd\") pod \"frr-k8s-webhook-server-64bf5d555-wcc2g\" (UID: \"6dd5ad64-5f6d-4e92-9c1b-3f0837cb134c\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wcc2g" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.705710 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f-metrics-certs\") pod \"frr-k8s-sv4vw\" (UID: \"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f\") " pod="metallb-system/frr-k8s-sv4vw" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.705738 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f-frr-startup\") pod \"frr-k8s-sv4vw\" (UID: \"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f\") " pod="metallb-system/frr-k8s-sv4vw" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.705761 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c9ac1753-f8ad-4f22-a955-38663b8121f0-metallb-excludel2\") pod \"speaker-bb2tj\" (UID: \"c9ac1753-f8ad-4f22-a955-38663b8121f0\") " pod="metallb-system/speaker-bb2tj" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.705759 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f-frr-conf\") pod \"frr-k8s-sv4vw\" (UID: \"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f\") " pod="metallb-system/frr-k8s-sv4vw" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.705789 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c9ac1753-f8ad-4f22-a955-38663b8121f0-memberlist\") pod \"speaker-bb2tj\" (UID: \"c9ac1753-f8ad-4f22-a955-38663b8121f0\") " pod="metallb-system/speaker-bb2tj" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.705871 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f-reloader\") pod \"frr-k8s-sv4vw\" (UID: \"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f\") " pod="metallb-system/frr-k8s-sv4vw" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.705867 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f-metrics\") pod \"frr-k8s-sv4vw\" (UID: \"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f\") " pod="metallb-system/frr-k8s-sv4vw" Oct 07 13:58:10 crc kubenswrapper[4959]: E1007 13:58:10.705971 4959 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Oct 07 13:58:10 crc kubenswrapper[4959]: E1007 13:58:10.706048 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f-metrics-certs podName:5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f nodeName:}" failed. No retries permitted until 2025-10-07 13:58:11.206025707 +0000 UTC m=+773.289430122 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f-metrics-certs") pod "frr-k8s-sv4vw" (UID: "5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f") : secret "frr-k8s-certs-secret" not found Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.706169 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f-reloader\") pod \"frr-k8s-sv4vw\" (UID: \"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f\") " pod="metallb-system/frr-k8s-sv4vw" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.706328 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f-frr-sockets\") pod \"frr-k8s-sv4vw\" (UID: \"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f\") " pod="metallb-system/frr-k8s-sv4vw" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.707136 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f-frr-startup\") pod \"frr-k8s-sv4vw\" (UID: \"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f\") " pod="metallb-system/frr-k8s-sv4vw" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.710710 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6dd5ad64-5f6d-4e92-9c1b-3f0837cb134c-cert\") pod \"frr-k8s-webhook-server-64bf5d555-wcc2g\" (UID: \"6dd5ad64-5f6d-4e92-9c1b-3f0837cb134c\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wcc2g" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.722580 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwddd\" (UniqueName: \"kubernetes.io/projected/6dd5ad64-5f6d-4e92-9c1b-3f0837cb134c-kube-api-access-nwddd\") pod \"frr-k8s-webhook-server-64bf5d555-wcc2g\" (UID: \"6dd5ad64-5f6d-4e92-9c1b-3f0837cb134c\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wcc2g" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.739523 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zl8q\" (UniqueName: \"kubernetes.io/projected/5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f-kube-api-access-8zl8q\") pod \"frr-k8s-sv4vw\" (UID: \"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f\") " pod="metallb-system/frr-k8s-sv4vw" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.807160 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c9ac1753-f8ad-4f22-a955-38663b8121f0-memberlist\") pod \"speaker-bb2tj\" (UID: \"c9ac1753-f8ad-4f22-a955-38663b8121f0\") " pod="metallb-system/speaker-bb2tj" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.807224 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mq7wl\" (UniqueName: \"kubernetes.io/projected/c9ac1753-f8ad-4f22-a955-38663b8121f0-kube-api-access-mq7wl\") pod \"speaker-bb2tj\" (UID: \"c9ac1753-f8ad-4f22-a955-38663b8121f0\") " pod="metallb-system/speaker-bb2tj" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.807255 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e2e00fa0-bc56-4a3f-9796-bbc1ec5dc47d-cert\") pod \"controller-68d546b9d8-nbdf7\" (UID: \"e2e00fa0-bc56-4a3f-9796-bbc1ec5dc47d\") " pod="metallb-system/controller-68d546b9d8-nbdf7" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.807279 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49hvz\" (UniqueName: \"kubernetes.io/projected/e2e00fa0-bc56-4a3f-9796-bbc1ec5dc47d-kube-api-access-49hvz\") pod \"controller-68d546b9d8-nbdf7\" (UID: \"e2e00fa0-bc56-4a3f-9796-bbc1ec5dc47d\") " pod="metallb-system/controller-68d546b9d8-nbdf7" Oct 07 13:58:10 crc kubenswrapper[4959]: E1007 13:58:10.807308 4959 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 07 13:58:10 crc kubenswrapper[4959]: E1007 13:58:10.807361 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c9ac1753-f8ad-4f22-a955-38663b8121f0-memberlist podName:c9ac1753-f8ad-4f22-a955-38663b8121f0 nodeName:}" failed. No retries permitted until 2025-10-07 13:58:11.30734609 +0000 UTC m=+773.390750415 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/c9ac1753-f8ad-4f22-a955-38663b8121f0-memberlist") pod "speaker-bb2tj" (UID: "c9ac1753-f8ad-4f22-a955-38663b8121f0") : secret "metallb-memberlist" not found Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.807311 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e2e00fa0-bc56-4a3f-9796-bbc1ec5dc47d-metrics-certs\") pod \"controller-68d546b9d8-nbdf7\" (UID: \"e2e00fa0-bc56-4a3f-9796-bbc1ec5dc47d\") " pod="metallb-system/controller-68d546b9d8-nbdf7" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.807409 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c9ac1753-f8ad-4f22-a955-38663b8121f0-metrics-certs\") pod \"speaker-bb2tj\" (UID: \"c9ac1753-f8ad-4f22-a955-38663b8121f0\") " pod="metallb-system/speaker-bb2tj" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.807468 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c9ac1753-f8ad-4f22-a955-38663b8121f0-metallb-excludel2\") pod \"speaker-bb2tj\" (UID: \"c9ac1753-f8ad-4f22-a955-38663b8121f0\") " pod="metallb-system/speaker-bb2tj" Oct 07 13:58:10 crc kubenswrapper[4959]: E1007 13:58:10.807720 4959 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Oct 07 13:58:10 crc kubenswrapper[4959]: E1007 13:58:10.807770 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c9ac1753-f8ad-4f22-a955-38663b8121f0-metrics-certs podName:c9ac1753-f8ad-4f22-a955-38663b8121f0 nodeName:}" failed. No retries permitted until 2025-10-07 13:58:11.30775624 +0000 UTC m=+773.391160565 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c9ac1753-f8ad-4f22-a955-38663b8121f0-metrics-certs") pod "speaker-bb2tj" (UID: "c9ac1753-f8ad-4f22-a955-38663b8121f0") : secret "speaker-certs-secret" not found Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.808060 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c9ac1753-f8ad-4f22-a955-38663b8121f0-metallb-excludel2\") pod \"speaker-bb2tj\" (UID: \"c9ac1753-f8ad-4f22-a955-38663b8121f0\") " pod="metallb-system/speaker-bb2tj" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.811531 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e2e00fa0-bc56-4a3f-9796-bbc1ec5dc47d-cert\") pod \"controller-68d546b9d8-nbdf7\" (UID: \"e2e00fa0-bc56-4a3f-9796-bbc1ec5dc47d\") " pod="metallb-system/controller-68d546b9d8-nbdf7" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.811840 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e2e00fa0-bc56-4a3f-9796-bbc1ec5dc47d-metrics-certs\") pod \"controller-68d546b9d8-nbdf7\" (UID: \"e2e00fa0-bc56-4a3f-9796-bbc1ec5dc47d\") " pod="metallb-system/controller-68d546b9d8-nbdf7" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.825094 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49hvz\" (UniqueName: \"kubernetes.io/projected/e2e00fa0-bc56-4a3f-9796-bbc1ec5dc47d-kube-api-access-49hvz\") pod \"controller-68d546b9d8-nbdf7\" (UID: \"e2e00fa0-bc56-4a3f-9796-bbc1ec5dc47d\") " pod="metallb-system/controller-68d546b9d8-nbdf7" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.829214 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mq7wl\" (UniqueName: \"kubernetes.io/projected/c9ac1753-f8ad-4f22-a955-38663b8121f0-kube-api-access-mq7wl\") pod \"speaker-bb2tj\" (UID: \"c9ac1753-f8ad-4f22-a955-38663b8121f0\") " pod="metallb-system/speaker-bb2tj" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.837600 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wcc2g" Oct 07 13:58:10 crc kubenswrapper[4959]: I1007 13:58:10.949399 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-nbdf7" Oct 07 13:58:11 crc kubenswrapper[4959]: I1007 13:58:11.215611 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f-metrics-certs\") pod \"frr-k8s-sv4vw\" (UID: \"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f\") " pod="metallb-system/frr-k8s-sv4vw" Oct 07 13:58:11 crc kubenswrapper[4959]: I1007 13:58:11.220028 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f-metrics-certs\") pod \"frr-k8s-sv4vw\" (UID: \"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f\") " pod="metallb-system/frr-k8s-sv4vw" Oct 07 13:58:11 crc kubenswrapper[4959]: I1007 13:58:11.252644 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-wcc2g"] Oct 07 13:58:11 crc kubenswrapper[4959]: I1007 13:58:11.317898 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c9ac1753-f8ad-4f22-a955-38663b8121f0-metrics-certs\") pod \"speaker-bb2tj\" (UID: \"c9ac1753-f8ad-4f22-a955-38663b8121f0\") " pod="metallb-system/speaker-bb2tj" Oct 07 13:58:11 crc kubenswrapper[4959]: I1007 13:58:11.317960 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c9ac1753-f8ad-4f22-a955-38663b8121f0-memberlist\") pod \"speaker-bb2tj\" (UID: \"c9ac1753-f8ad-4f22-a955-38663b8121f0\") " pod="metallb-system/speaker-bb2tj" Oct 07 13:58:11 crc kubenswrapper[4959]: E1007 13:58:11.318079 4959 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 07 13:58:11 crc kubenswrapper[4959]: E1007 13:58:11.318137 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c9ac1753-f8ad-4f22-a955-38663b8121f0-memberlist podName:c9ac1753-f8ad-4f22-a955-38663b8121f0 nodeName:}" failed. No retries permitted until 2025-10-07 13:58:12.318124638 +0000 UTC m=+774.401528963 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/c9ac1753-f8ad-4f22-a955-38663b8121f0-memberlist") pod "speaker-bb2tj" (UID: "c9ac1753-f8ad-4f22-a955-38663b8121f0") : secret "metallb-memberlist" not found Oct 07 13:58:11 crc kubenswrapper[4959]: I1007 13:58:11.322661 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c9ac1753-f8ad-4f22-a955-38663b8121f0-metrics-certs\") pod \"speaker-bb2tj\" (UID: \"c9ac1753-f8ad-4f22-a955-38663b8121f0\") " pod="metallb-system/speaker-bb2tj" Oct 07 13:58:11 crc kubenswrapper[4959]: I1007 13:58:11.364977 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-nbdf7"] Oct 07 13:58:11 crc kubenswrapper[4959]: W1007 13:58:11.374072 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode2e00fa0_bc56_4a3f_9796_bbc1ec5dc47d.slice/crio-1a66988492783212a9ba1ade71c5304d158ba3231dcd645ffc2195db3f552bab WatchSource:0}: Error finding container 1a66988492783212a9ba1ade71c5304d158ba3231dcd645ffc2195db3f552bab: Status 404 returned error can't find the container with id 1a66988492783212a9ba1ade71c5304d158ba3231dcd645ffc2195db3f552bab Oct 07 13:58:11 crc kubenswrapper[4959]: I1007 13:58:11.427683 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-sv4vw" Oct 07 13:58:11 crc kubenswrapper[4959]: I1007 13:58:11.647521 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sv4vw" event={"ID":"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f","Type":"ContainerStarted","Data":"83cc1e039b5ed6bc7afe05e23eee1ec9368c8588566bf606f7daeaec0dedb169"} Oct 07 13:58:11 crc kubenswrapper[4959]: I1007 13:58:11.648347 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wcc2g" event={"ID":"6dd5ad64-5f6d-4e92-9c1b-3f0837cb134c","Type":"ContainerStarted","Data":"0e5fcbf11dc260c94d6cf0aa47301600c48ff20b4e9633f44ad1bbb6b608ec42"} Oct 07 13:58:11 crc kubenswrapper[4959]: I1007 13:58:11.649929 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-nbdf7" event={"ID":"e2e00fa0-bc56-4a3f-9796-bbc1ec5dc47d","Type":"ContainerStarted","Data":"36dac7c2e343cb73614b2598f5fb16cc6c9c2a7dde1ada96163125df7b0de136"} Oct 07 13:58:11 crc kubenswrapper[4959]: I1007 13:58:11.649953 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-nbdf7" event={"ID":"e2e00fa0-bc56-4a3f-9796-bbc1ec5dc47d","Type":"ContainerStarted","Data":"acd455309a3fe655c5ab2e47e771af5aad136fb0f4dfd70ed76903bc21d01060"} Oct 07 13:58:11 crc kubenswrapper[4959]: I1007 13:58:11.649966 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-nbdf7" event={"ID":"e2e00fa0-bc56-4a3f-9796-bbc1ec5dc47d","Type":"ContainerStarted","Data":"1a66988492783212a9ba1ade71c5304d158ba3231dcd645ffc2195db3f552bab"} Oct 07 13:58:11 crc kubenswrapper[4959]: I1007 13:58:11.650193 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-nbdf7" Oct 07 13:58:11 crc kubenswrapper[4959]: I1007 13:58:11.673557 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-nbdf7" podStartSLOduration=1.673535286 podStartE2EDuration="1.673535286s" podCreationTimestamp="2025-10-07 13:58:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:58:11.669868403 +0000 UTC m=+773.753272738" watchObservedRunningTime="2025-10-07 13:58:11.673535286 +0000 UTC m=+773.756939611" Oct 07 13:58:12 crc kubenswrapper[4959]: I1007 13:58:12.330024 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c9ac1753-f8ad-4f22-a955-38663b8121f0-memberlist\") pod \"speaker-bb2tj\" (UID: \"c9ac1753-f8ad-4f22-a955-38663b8121f0\") " pod="metallb-system/speaker-bb2tj" Oct 07 13:58:12 crc kubenswrapper[4959]: I1007 13:58:12.352764 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c9ac1753-f8ad-4f22-a955-38663b8121f0-memberlist\") pod \"speaker-bb2tj\" (UID: \"c9ac1753-f8ad-4f22-a955-38663b8121f0\") " pod="metallb-system/speaker-bb2tj" Oct 07 13:58:12 crc kubenswrapper[4959]: I1007 13:58:12.428954 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-bb2tj" Oct 07 13:58:12 crc kubenswrapper[4959]: W1007 13:58:12.447422 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc9ac1753_f8ad_4f22_a955_38663b8121f0.slice/crio-9d375295a2ec37053bc165f10fffe4884b98cc340d5c5276101b2e7dee9be348 WatchSource:0}: Error finding container 9d375295a2ec37053bc165f10fffe4884b98cc340d5c5276101b2e7dee9be348: Status 404 returned error can't find the container with id 9d375295a2ec37053bc165f10fffe4884b98cc340d5c5276101b2e7dee9be348 Oct 07 13:58:12 crc kubenswrapper[4959]: I1007 13:58:12.664480 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bb2tj" event={"ID":"c9ac1753-f8ad-4f22-a955-38663b8121f0","Type":"ContainerStarted","Data":"9d375295a2ec37053bc165f10fffe4884b98cc340d5c5276101b2e7dee9be348"} Oct 07 13:58:13 crc kubenswrapper[4959]: I1007 13:58:13.672791 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bb2tj" event={"ID":"c9ac1753-f8ad-4f22-a955-38663b8121f0","Type":"ContainerStarted","Data":"9154f1c02f81dec141b9c88844662c98e60349105d88d5a3f2dc9caa1852cc35"} Oct 07 13:58:13 crc kubenswrapper[4959]: I1007 13:58:13.673149 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bb2tj" event={"ID":"c9ac1753-f8ad-4f22-a955-38663b8121f0","Type":"ContainerStarted","Data":"af53655f9b7a35cc9e5051a7706fccad0d63585c70bf6097d6e38f2f8be5e098"} Oct 07 13:58:13 crc kubenswrapper[4959]: I1007 13:58:13.673453 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-bb2tj" Oct 07 13:58:13 crc kubenswrapper[4959]: I1007 13:58:13.690891 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-bb2tj" podStartSLOduration=3.690871515 podStartE2EDuration="3.690871515s" podCreationTimestamp="2025-10-07 13:58:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:58:13.686518055 +0000 UTC m=+775.769922380" watchObservedRunningTime="2025-10-07 13:58:13.690871515 +0000 UTC m=+775.774275840" Oct 07 13:58:18 crc kubenswrapper[4959]: I1007 13:58:18.713618 4959 generic.go:334] "Generic (PLEG): container finished" podID="5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f" containerID="78752617027da938cc929b03f77e1e324bd4f02877a5df6b7dd21e3d3e94b7ec" exitCode=0 Oct 07 13:58:18 crc kubenswrapper[4959]: I1007 13:58:18.714610 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sv4vw" event={"ID":"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f","Type":"ContainerDied","Data":"78752617027da938cc929b03f77e1e324bd4f02877a5df6b7dd21e3d3e94b7ec"} Oct 07 13:58:18 crc kubenswrapper[4959]: I1007 13:58:18.716831 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wcc2g" event={"ID":"6dd5ad64-5f6d-4e92-9c1b-3f0837cb134c","Type":"ContainerStarted","Data":"1f8f2985deefe23c09474f7dd31803134b51f3acd13a2414bad29ec5f2c2aa97"} Oct 07 13:58:18 crc kubenswrapper[4959]: I1007 13:58:18.717093 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wcc2g" Oct 07 13:58:18 crc kubenswrapper[4959]: I1007 13:58:18.769771 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wcc2g" podStartSLOduration=2.124718357 podStartE2EDuration="8.769744491s" podCreationTimestamp="2025-10-07 13:58:10 +0000 UTC" firstStartedPulling="2025-10-07 13:58:11.261861475 +0000 UTC m=+773.345265800" lastFinishedPulling="2025-10-07 13:58:17.906887609 +0000 UTC m=+779.990291934" observedRunningTime="2025-10-07 13:58:18.759338455 +0000 UTC m=+780.842742800" watchObservedRunningTime="2025-10-07 13:58:18.769744491 +0000 UTC m=+780.853148826" Oct 07 13:58:19 crc kubenswrapper[4959]: I1007 13:58:19.726974 4959 generic.go:334] "Generic (PLEG): container finished" podID="5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f" containerID="9e970200756d31fc8ff2aecbbf349eaa7d193d88062009ac9e2d242223b95636" exitCode=0 Oct 07 13:58:19 crc kubenswrapper[4959]: I1007 13:58:19.727092 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sv4vw" event={"ID":"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f","Type":"ContainerDied","Data":"9e970200756d31fc8ff2aecbbf349eaa7d193d88062009ac9e2d242223b95636"} Oct 07 13:58:20 crc kubenswrapper[4959]: I1007 13:58:20.630784 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 13:58:20 crc kubenswrapper[4959]: I1007 13:58:20.630852 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 13:58:20 crc kubenswrapper[4959]: I1007 13:58:20.630904 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 13:58:20 crc kubenswrapper[4959]: I1007 13:58:20.631654 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"59698f7bbe675b82b270fb68b1462a80d17b9a12f2121c8b8aa53d105e879f24"} pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 13:58:20 crc kubenswrapper[4959]: I1007 13:58:20.631725 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" containerID="cri-o://59698f7bbe675b82b270fb68b1462a80d17b9a12f2121c8b8aa53d105e879f24" gracePeriod=600 Oct 07 13:58:20 crc kubenswrapper[4959]: I1007 13:58:20.735753 4959 generic.go:334] "Generic (PLEG): container finished" podID="5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f" containerID="e95a31a403cb7ce3e0abe0296ace22c6beb57d297962fd72da91403038290fc8" exitCode=0 Oct 07 13:58:20 crc kubenswrapper[4959]: I1007 13:58:20.735838 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sv4vw" event={"ID":"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f","Type":"ContainerDied","Data":"e95a31a403cb7ce3e0abe0296ace22c6beb57d297962fd72da91403038290fc8"} Oct 07 13:58:21 crc kubenswrapper[4959]: I1007 13:58:21.757997 4959 generic.go:334] "Generic (PLEG): container finished" podID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerID="59698f7bbe675b82b270fb68b1462a80d17b9a12f2121c8b8aa53d105e879f24" exitCode=0 Oct 07 13:58:21 crc kubenswrapper[4959]: I1007 13:58:21.758149 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerDied","Data":"59698f7bbe675b82b270fb68b1462a80d17b9a12f2121c8b8aa53d105e879f24"} Oct 07 13:58:21 crc kubenswrapper[4959]: I1007 13:58:21.758784 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerStarted","Data":"be8cac39c0f211821bf3cae10b6e2f58082165f269df748f76ccd66a0c6a1240"} Oct 07 13:58:21 crc kubenswrapper[4959]: I1007 13:58:21.758812 4959 scope.go:117] "RemoveContainer" containerID="e4d224bb95cdc421b51fa1f72abe6b70cf3c87a3b06fde77476b65c5f022b080" Oct 07 13:58:21 crc kubenswrapper[4959]: I1007 13:58:21.765711 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sv4vw" event={"ID":"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f","Type":"ContainerStarted","Data":"e4b24f5a105557a4ca3582f1c9f82a2529ffee840702a76511a180243c4b77de"} Oct 07 13:58:21 crc kubenswrapper[4959]: I1007 13:58:21.765756 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sv4vw" event={"ID":"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f","Type":"ContainerStarted","Data":"8bdd6caa6c42d9967efeef251aefc6a2c3104c522c471f640579a72bb3b9f7e7"} Oct 07 13:58:21 crc kubenswrapper[4959]: I1007 13:58:21.765772 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sv4vw" event={"ID":"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f","Type":"ContainerStarted","Data":"99f5a156791819122364bdc7e771b14582b70c448a6f86e5c55bef96a4d880a2"} Oct 07 13:58:21 crc kubenswrapper[4959]: I1007 13:58:21.765783 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sv4vw" event={"ID":"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f","Type":"ContainerStarted","Data":"fec1dbb1af1e86a2225a78b60a2d5122356975d66b139ca80c743500c99ee727"} Oct 07 13:58:21 crc kubenswrapper[4959]: I1007 13:58:21.765795 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sv4vw" event={"ID":"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f","Type":"ContainerStarted","Data":"11f19c454b5108672f6177caae260a0a3ed80032e45b0cd7b5e825fa1f1d0da3"} Oct 07 13:58:22 crc kubenswrapper[4959]: I1007 13:58:22.433608 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-bb2tj" Oct 07 13:58:22 crc kubenswrapper[4959]: I1007 13:58:22.780569 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sv4vw" event={"ID":"5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f","Type":"ContainerStarted","Data":"5585da0c238b80c18f805cb04c5616b8c504d4e667b0389b040b3384e77250ae"} Oct 07 13:58:22 crc kubenswrapper[4959]: I1007 13:58:22.781118 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-sv4vw" Oct 07 13:58:22 crc kubenswrapper[4959]: I1007 13:58:22.810709 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-sv4vw" podStartSLOduration=6.38619845 podStartE2EDuration="12.810692722s" podCreationTimestamp="2025-10-07 13:58:10 +0000 UTC" firstStartedPulling="2025-10-07 13:58:11.503894326 +0000 UTC m=+773.587298641" lastFinishedPulling="2025-10-07 13:58:17.928388568 +0000 UTC m=+780.011792913" observedRunningTime="2025-10-07 13:58:22.808221981 +0000 UTC m=+784.891626336" watchObservedRunningTime="2025-10-07 13:58:22.810692722 +0000 UTC m=+784.894097047" Oct 07 13:58:25 crc kubenswrapper[4959]: I1007 13:58:25.540312 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-stg4g"] Oct 07 13:58:25 crc kubenswrapper[4959]: I1007 13:58:25.541460 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-stg4g" Oct 07 13:58:25 crc kubenswrapper[4959]: I1007 13:58:25.544601 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-l9c4b" Oct 07 13:58:25 crc kubenswrapper[4959]: I1007 13:58:25.547158 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 07 13:58:25 crc kubenswrapper[4959]: I1007 13:58:25.547195 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 07 13:58:25 crc kubenswrapper[4959]: I1007 13:58:25.565388 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-stg4g"] Oct 07 13:58:25 crc kubenswrapper[4959]: I1007 13:58:25.616658 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mx2nk\" (UniqueName: \"kubernetes.io/projected/62231a9a-c3c1-46a8-90f2-d5c7775eabc8-kube-api-access-mx2nk\") pod \"openstack-operator-index-stg4g\" (UID: \"62231a9a-c3c1-46a8-90f2-d5c7775eabc8\") " pod="openstack-operators/openstack-operator-index-stg4g" Oct 07 13:58:25 crc kubenswrapper[4959]: I1007 13:58:25.717403 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mx2nk\" (UniqueName: \"kubernetes.io/projected/62231a9a-c3c1-46a8-90f2-d5c7775eabc8-kube-api-access-mx2nk\") pod \"openstack-operator-index-stg4g\" (UID: \"62231a9a-c3c1-46a8-90f2-d5c7775eabc8\") " pod="openstack-operators/openstack-operator-index-stg4g" Oct 07 13:58:25 crc kubenswrapper[4959]: I1007 13:58:25.736127 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mx2nk\" (UniqueName: \"kubernetes.io/projected/62231a9a-c3c1-46a8-90f2-d5c7775eabc8-kube-api-access-mx2nk\") pod \"openstack-operator-index-stg4g\" (UID: \"62231a9a-c3c1-46a8-90f2-d5c7775eabc8\") " pod="openstack-operators/openstack-operator-index-stg4g" Oct 07 13:58:25 crc kubenswrapper[4959]: I1007 13:58:25.870248 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-stg4g" Oct 07 13:58:26 crc kubenswrapper[4959]: I1007 13:58:26.380262 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-stg4g"] Oct 07 13:58:26 crc kubenswrapper[4959]: I1007 13:58:26.428133 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-sv4vw" Oct 07 13:58:26 crc kubenswrapper[4959]: I1007 13:58:26.479763 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-sv4vw" Oct 07 13:58:26 crc kubenswrapper[4959]: I1007 13:58:26.801604 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-stg4g" event={"ID":"62231a9a-c3c1-46a8-90f2-d5c7775eabc8","Type":"ContainerStarted","Data":"caf1bce5ea2e0c8a6953c6f4449e3d73ae24c0e2e1c33ffbbfcc19e55816c257"} Oct 07 13:58:27 crc kubenswrapper[4959]: I1007 13:58:27.810270 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-stg4g" event={"ID":"62231a9a-c3c1-46a8-90f2-d5c7775eabc8","Type":"ContainerStarted","Data":"5cd1fea1b9db1d8437c4ed4d0a1f9dcc8d97a01c602c3ec17202db68190ecfb4"} Oct 07 13:58:27 crc kubenswrapper[4959]: I1007 13:58:27.840000 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-stg4g" podStartSLOduration=1.911611588 podStartE2EDuration="2.839961831s" podCreationTimestamp="2025-10-07 13:58:25 +0000 UTC" firstStartedPulling="2025-10-07 13:58:26.386445546 +0000 UTC m=+788.469849911" lastFinishedPulling="2025-10-07 13:58:27.314795829 +0000 UTC m=+789.398200154" observedRunningTime="2025-10-07 13:58:27.833210665 +0000 UTC m=+789.916615030" watchObservedRunningTime="2025-10-07 13:58:27.839961831 +0000 UTC m=+789.923366196" Oct 07 13:58:28 crc kubenswrapper[4959]: I1007 13:58:28.529685 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-stg4g"] Oct 07 13:58:29 crc kubenswrapper[4959]: I1007 13:58:29.140056 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-28476"] Oct 07 13:58:29 crc kubenswrapper[4959]: I1007 13:58:29.141692 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-28476" Oct 07 13:58:29 crc kubenswrapper[4959]: I1007 13:58:29.147049 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-28476"] Oct 07 13:58:29 crc kubenswrapper[4959]: I1007 13:58:29.268284 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mcjv\" (UniqueName: \"kubernetes.io/projected/51c0a7d2-8289-4c41-b419-16db5b361e72-kube-api-access-5mcjv\") pod \"openstack-operator-index-28476\" (UID: \"51c0a7d2-8289-4c41-b419-16db5b361e72\") " pod="openstack-operators/openstack-operator-index-28476" Oct 07 13:58:29 crc kubenswrapper[4959]: I1007 13:58:29.370333 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mcjv\" (UniqueName: \"kubernetes.io/projected/51c0a7d2-8289-4c41-b419-16db5b361e72-kube-api-access-5mcjv\") pod \"openstack-operator-index-28476\" (UID: \"51c0a7d2-8289-4c41-b419-16db5b361e72\") " pod="openstack-operators/openstack-operator-index-28476" Oct 07 13:58:29 crc kubenswrapper[4959]: I1007 13:58:29.393201 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mcjv\" (UniqueName: \"kubernetes.io/projected/51c0a7d2-8289-4c41-b419-16db5b361e72-kube-api-access-5mcjv\") pod \"openstack-operator-index-28476\" (UID: \"51c0a7d2-8289-4c41-b419-16db5b361e72\") " pod="openstack-operators/openstack-operator-index-28476" Oct 07 13:58:29 crc kubenswrapper[4959]: I1007 13:58:29.467003 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-28476" Oct 07 13:58:29 crc kubenswrapper[4959]: I1007 13:58:29.824703 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-stg4g" podUID="62231a9a-c3c1-46a8-90f2-d5c7775eabc8" containerName="registry-server" containerID="cri-o://5cd1fea1b9db1d8437c4ed4d0a1f9dcc8d97a01c602c3ec17202db68190ecfb4" gracePeriod=2 Oct 07 13:58:29 crc kubenswrapper[4959]: I1007 13:58:29.900080 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-28476"] Oct 07 13:58:29 crc kubenswrapper[4959]: W1007 13:58:29.953157 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod51c0a7d2_8289_4c41_b419_16db5b361e72.slice/crio-5cee51ed2a18c29d3fdf62790bf14d38872f7cc2e633b685b062de5b4c5c1662 WatchSource:0}: Error finding container 5cee51ed2a18c29d3fdf62790bf14d38872f7cc2e633b685b062de5b4c5c1662: Status 404 returned error can't find the container with id 5cee51ed2a18c29d3fdf62790bf14d38872f7cc2e633b685b062de5b4c5c1662 Oct 07 13:58:30 crc kubenswrapper[4959]: I1007 13:58:30.208827 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-stg4g" Oct 07 13:58:30 crc kubenswrapper[4959]: I1007 13:58:30.281612 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mx2nk\" (UniqueName: \"kubernetes.io/projected/62231a9a-c3c1-46a8-90f2-d5c7775eabc8-kube-api-access-mx2nk\") pod \"62231a9a-c3c1-46a8-90f2-d5c7775eabc8\" (UID: \"62231a9a-c3c1-46a8-90f2-d5c7775eabc8\") " Oct 07 13:58:30 crc kubenswrapper[4959]: I1007 13:58:30.287667 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62231a9a-c3c1-46a8-90f2-d5c7775eabc8-kube-api-access-mx2nk" (OuterVolumeSpecName: "kube-api-access-mx2nk") pod "62231a9a-c3c1-46a8-90f2-d5c7775eabc8" (UID: "62231a9a-c3c1-46a8-90f2-d5c7775eabc8"). InnerVolumeSpecName "kube-api-access-mx2nk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:58:30 crc kubenswrapper[4959]: I1007 13:58:30.382875 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mx2nk\" (UniqueName: \"kubernetes.io/projected/62231a9a-c3c1-46a8-90f2-d5c7775eabc8-kube-api-access-mx2nk\") on node \"crc\" DevicePath \"\"" Oct 07 13:58:30 crc kubenswrapper[4959]: I1007 13:58:30.833111 4959 generic.go:334] "Generic (PLEG): container finished" podID="62231a9a-c3c1-46a8-90f2-d5c7775eabc8" containerID="5cd1fea1b9db1d8437c4ed4d0a1f9dcc8d97a01c602c3ec17202db68190ecfb4" exitCode=0 Oct 07 13:58:30 crc kubenswrapper[4959]: I1007 13:58:30.833251 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-stg4g" event={"ID":"62231a9a-c3c1-46a8-90f2-d5c7775eabc8","Type":"ContainerDied","Data":"5cd1fea1b9db1d8437c4ed4d0a1f9dcc8d97a01c602c3ec17202db68190ecfb4"} Oct 07 13:58:30 crc kubenswrapper[4959]: I1007 13:58:30.833483 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-stg4g" event={"ID":"62231a9a-c3c1-46a8-90f2-d5c7775eabc8","Type":"ContainerDied","Data":"caf1bce5ea2e0c8a6953c6f4449e3d73ae24c0e2e1c33ffbbfcc19e55816c257"} Oct 07 13:58:30 crc kubenswrapper[4959]: I1007 13:58:30.833503 4959 scope.go:117] "RemoveContainer" containerID="5cd1fea1b9db1d8437c4ed4d0a1f9dcc8d97a01c602c3ec17202db68190ecfb4" Oct 07 13:58:30 crc kubenswrapper[4959]: I1007 13:58:30.833344 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-stg4g" Oct 07 13:58:30 crc kubenswrapper[4959]: I1007 13:58:30.839978 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-28476" event={"ID":"51c0a7d2-8289-4c41-b419-16db5b361e72","Type":"ContainerStarted","Data":"18432719db9ce5f8afeaf3ca419c1d8a7e002672fb6500857c5479da73811488"} Oct 07 13:58:30 crc kubenswrapper[4959]: I1007 13:58:30.840032 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-28476" event={"ID":"51c0a7d2-8289-4c41-b419-16db5b361e72","Type":"ContainerStarted","Data":"5cee51ed2a18c29d3fdf62790bf14d38872f7cc2e633b685b062de5b4c5c1662"} Oct 07 13:58:30 crc kubenswrapper[4959]: I1007 13:58:30.844001 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-wcc2g" Oct 07 13:58:30 crc kubenswrapper[4959]: I1007 13:58:30.862828 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-28476" podStartSLOduration=1.274488464 podStartE2EDuration="1.862530264s" podCreationTimestamp="2025-10-07 13:58:29 +0000 UTC" firstStartedPulling="2025-10-07 13:58:29.956732915 +0000 UTC m=+792.040137260" lastFinishedPulling="2025-10-07 13:58:30.544774685 +0000 UTC m=+792.628179060" observedRunningTime="2025-10-07 13:58:30.860142265 +0000 UTC m=+792.943546590" watchObservedRunningTime="2025-10-07 13:58:30.862530264 +0000 UTC m=+792.945934589" Oct 07 13:58:30 crc kubenswrapper[4959]: I1007 13:58:30.868647 4959 scope.go:117] "RemoveContainer" containerID="5cd1fea1b9db1d8437c4ed4d0a1f9dcc8d97a01c602c3ec17202db68190ecfb4" Oct 07 13:58:30 crc kubenswrapper[4959]: E1007 13:58:30.876769 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cd1fea1b9db1d8437c4ed4d0a1f9dcc8d97a01c602c3ec17202db68190ecfb4\": container with ID starting with 5cd1fea1b9db1d8437c4ed4d0a1f9dcc8d97a01c602c3ec17202db68190ecfb4 not found: ID does not exist" containerID="5cd1fea1b9db1d8437c4ed4d0a1f9dcc8d97a01c602c3ec17202db68190ecfb4" Oct 07 13:58:30 crc kubenswrapper[4959]: I1007 13:58:30.876828 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cd1fea1b9db1d8437c4ed4d0a1f9dcc8d97a01c602c3ec17202db68190ecfb4"} err="failed to get container status \"5cd1fea1b9db1d8437c4ed4d0a1f9dcc8d97a01c602c3ec17202db68190ecfb4\": rpc error: code = NotFound desc = could not find container \"5cd1fea1b9db1d8437c4ed4d0a1f9dcc8d97a01c602c3ec17202db68190ecfb4\": container with ID starting with 5cd1fea1b9db1d8437c4ed4d0a1f9dcc8d97a01c602c3ec17202db68190ecfb4 not found: ID does not exist" Oct 07 13:58:30 crc kubenswrapper[4959]: I1007 13:58:30.877312 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-stg4g"] Oct 07 13:58:30 crc kubenswrapper[4959]: I1007 13:58:30.880479 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-stg4g"] Oct 07 13:58:30 crc kubenswrapper[4959]: I1007 13:58:30.955866 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-nbdf7" Oct 07 13:58:31 crc kubenswrapper[4959]: I1007 13:58:31.431472 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-sv4vw" Oct 07 13:58:32 crc kubenswrapper[4959]: I1007 13:58:32.664562 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62231a9a-c3c1-46a8-90f2-d5c7775eabc8" path="/var/lib/kubelet/pods/62231a9a-c3c1-46a8-90f2-d5c7775eabc8/volumes" Oct 07 13:58:39 crc kubenswrapper[4959]: I1007 13:58:39.468059 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-28476" Oct 07 13:58:39 crc kubenswrapper[4959]: I1007 13:58:39.469218 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-28476" Oct 07 13:58:39 crc kubenswrapper[4959]: I1007 13:58:39.518856 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-28476" Oct 07 13:58:39 crc kubenswrapper[4959]: I1007 13:58:39.947958 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-28476" Oct 07 13:58:43 crc kubenswrapper[4959]: I1007 13:58:43.499296 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ljhpw"] Oct 07 13:58:43 crc kubenswrapper[4959]: E1007 13:58:43.500027 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62231a9a-c3c1-46a8-90f2-d5c7775eabc8" containerName="registry-server" Oct 07 13:58:43 crc kubenswrapper[4959]: I1007 13:58:43.500043 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="62231a9a-c3c1-46a8-90f2-d5c7775eabc8" containerName="registry-server" Oct 07 13:58:43 crc kubenswrapper[4959]: I1007 13:58:43.500241 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="62231a9a-c3c1-46a8-90f2-d5c7775eabc8" containerName="registry-server" Oct 07 13:58:43 crc kubenswrapper[4959]: I1007 13:58:43.501663 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ljhpw" Oct 07 13:58:43 crc kubenswrapper[4959]: I1007 13:58:43.517245 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ljhpw"] Oct 07 13:58:43 crc kubenswrapper[4959]: I1007 13:58:43.571602 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6380bfbe-af4a-4a1e-a61b-cf57fbae3e72-utilities\") pod \"community-operators-ljhpw\" (UID: \"6380bfbe-af4a-4a1e-a61b-cf57fbae3e72\") " pod="openshift-marketplace/community-operators-ljhpw" Oct 07 13:58:43 crc kubenswrapper[4959]: I1007 13:58:43.571687 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6380bfbe-af4a-4a1e-a61b-cf57fbae3e72-catalog-content\") pod \"community-operators-ljhpw\" (UID: \"6380bfbe-af4a-4a1e-a61b-cf57fbae3e72\") " pod="openshift-marketplace/community-operators-ljhpw" Oct 07 13:58:43 crc kubenswrapper[4959]: I1007 13:58:43.571748 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjffb\" (UniqueName: \"kubernetes.io/projected/6380bfbe-af4a-4a1e-a61b-cf57fbae3e72-kube-api-access-xjffb\") pod \"community-operators-ljhpw\" (UID: \"6380bfbe-af4a-4a1e-a61b-cf57fbae3e72\") " pod="openshift-marketplace/community-operators-ljhpw" Oct 07 13:58:43 crc kubenswrapper[4959]: I1007 13:58:43.673399 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6380bfbe-af4a-4a1e-a61b-cf57fbae3e72-utilities\") pod \"community-operators-ljhpw\" (UID: \"6380bfbe-af4a-4a1e-a61b-cf57fbae3e72\") " pod="openshift-marketplace/community-operators-ljhpw" Oct 07 13:58:43 crc kubenswrapper[4959]: I1007 13:58:43.673770 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6380bfbe-af4a-4a1e-a61b-cf57fbae3e72-catalog-content\") pod \"community-operators-ljhpw\" (UID: \"6380bfbe-af4a-4a1e-a61b-cf57fbae3e72\") " pod="openshift-marketplace/community-operators-ljhpw" Oct 07 13:58:43 crc kubenswrapper[4959]: I1007 13:58:43.673972 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjffb\" (UniqueName: \"kubernetes.io/projected/6380bfbe-af4a-4a1e-a61b-cf57fbae3e72-kube-api-access-xjffb\") pod \"community-operators-ljhpw\" (UID: \"6380bfbe-af4a-4a1e-a61b-cf57fbae3e72\") " pod="openshift-marketplace/community-operators-ljhpw" Oct 07 13:58:43 crc kubenswrapper[4959]: I1007 13:58:43.674308 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6380bfbe-af4a-4a1e-a61b-cf57fbae3e72-utilities\") pod \"community-operators-ljhpw\" (UID: \"6380bfbe-af4a-4a1e-a61b-cf57fbae3e72\") " pod="openshift-marketplace/community-operators-ljhpw" Oct 07 13:58:43 crc kubenswrapper[4959]: I1007 13:58:43.674448 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6380bfbe-af4a-4a1e-a61b-cf57fbae3e72-catalog-content\") pod \"community-operators-ljhpw\" (UID: \"6380bfbe-af4a-4a1e-a61b-cf57fbae3e72\") " pod="openshift-marketplace/community-operators-ljhpw" Oct 07 13:58:43 crc kubenswrapper[4959]: I1007 13:58:43.702558 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjffb\" (UniqueName: \"kubernetes.io/projected/6380bfbe-af4a-4a1e-a61b-cf57fbae3e72-kube-api-access-xjffb\") pod \"community-operators-ljhpw\" (UID: \"6380bfbe-af4a-4a1e-a61b-cf57fbae3e72\") " pod="openshift-marketplace/community-operators-ljhpw" Oct 07 13:58:43 crc kubenswrapper[4959]: I1007 13:58:43.825246 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ljhpw" Oct 07 13:58:44 crc kubenswrapper[4959]: W1007 13:58:44.253579 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6380bfbe_af4a_4a1e_a61b_cf57fbae3e72.slice/crio-c410e74664e9a36579ea5f0c70e87b6083a97e34a23d5d494ca9da7abf6a6e32 WatchSource:0}: Error finding container c410e74664e9a36579ea5f0c70e87b6083a97e34a23d5d494ca9da7abf6a6e32: Status 404 returned error can't find the container with id c410e74664e9a36579ea5f0c70e87b6083a97e34a23d5d494ca9da7abf6a6e32 Oct 07 13:58:44 crc kubenswrapper[4959]: I1007 13:58:44.253763 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ljhpw"] Oct 07 13:58:44 crc kubenswrapper[4959]: I1007 13:58:44.951194 4959 generic.go:334] "Generic (PLEG): container finished" podID="6380bfbe-af4a-4a1e-a61b-cf57fbae3e72" containerID="c0e7dbeac848a5694353af199602ae125a49d3cd4f10e63aa814ed5306a05262" exitCode=0 Oct 07 13:58:44 crc kubenswrapper[4959]: I1007 13:58:44.951283 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ljhpw" event={"ID":"6380bfbe-af4a-4a1e-a61b-cf57fbae3e72","Type":"ContainerDied","Data":"c0e7dbeac848a5694353af199602ae125a49d3cd4f10e63aa814ed5306a05262"} Oct 07 13:58:44 crc kubenswrapper[4959]: I1007 13:58:44.951606 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ljhpw" event={"ID":"6380bfbe-af4a-4a1e-a61b-cf57fbae3e72","Type":"ContainerStarted","Data":"c410e74664e9a36579ea5f0c70e87b6083a97e34a23d5d494ca9da7abf6a6e32"} Oct 07 13:58:46 crc kubenswrapper[4959]: I1007 13:58:46.514820 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv"] Oct 07 13:58:46 crc kubenswrapper[4959]: I1007 13:58:46.516497 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv" Oct 07 13:58:46 crc kubenswrapper[4959]: I1007 13:58:46.517879 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-6flnl" Oct 07 13:58:46 crc kubenswrapper[4959]: I1007 13:58:46.528188 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv"] Oct 07 13:58:46 crc kubenswrapper[4959]: I1007 13:58:46.544259 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/209faffc-bcb0-4a9b-8558-eab623e1b675-bundle\") pod \"563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv\" (UID: \"209faffc-bcb0-4a9b-8558-eab623e1b675\") " pod="openstack-operators/563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv" Oct 07 13:58:46 crc kubenswrapper[4959]: I1007 13:58:46.544318 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/209faffc-bcb0-4a9b-8558-eab623e1b675-util\") pod \"563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv\" (UID: \"209faffc-bcb0-4a9b-8558-eab623e1b675\") " pod="openstack-operators/563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv" Oct 07 13:58:46 crc kubenswrapper[4959]: I1007 13:58:46.544354 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcqcb\" (UniqueName: \"kubernetes.io/projected/209faffc-bcb0-4a9b-8558-eab623e1b675-kube-api-access-xcqcb\") pod \"563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv\" (UID: \"209faffc-bcb0-4a9b-8558-eab623e1b675\") " pod="openstack-operators/563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv" Oct 07 13:58:46 crc kubenswrapper[4959]: I1007 13:58:46.645714 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/209faffc-bcb0-4a9b-8558-eab623e1b675-util\") pod \"563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv\" (UID: \"209faffc-bcb0-4a9b-8558-eab623e1b675\") " pod="openstack-operators/563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv" Oct 07 13:58:46 crc kubenswrapper[4959]: I1007 13:58:46.645807 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcqcb\" (UniqueName: \"kubernetes.io/projected/209faffc-bcb0-4a9b-8558-eab623e1b675-kube-api-access-xcqcb\") pod \"563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv\" (UID: \"209faffc-bcb0-4a9b-8558-eab623e1b675\") " pod="openstack-operators/563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv" Oct 07 13:58:46 crc kubenswrapper[4959]: I1007 13:58:46.645921 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/209faffc-bcb0-4a9b-8558-eab623e1b675-bundle\") pod \"563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv\" (UID: \"209faffc-bcb0-4a9b-8558-eab623e1b675\") " pod="openstack-operators/563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv" Oct 07 13:58:46 crc kubenswrapper[4959]: I1007 13:58:46.646261 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/209faffc-bcb0-4a9b-8558-eab623e1b675-util\") pod \"563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv\" (UID: \"209faffc-bcb0-4a9b-8558-eab623e1b675\") " pod="openstack-operators/563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv" Oct 07 13:58:46 crc kubenswrapper[4959]: I1007 13:58:46.646461 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/209faffc-bcb0-4a9b-8558-eab623e1b675-bundle\") pod \"563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv\" (UID: \"209faffc-bcb0-4a9b-8558-eab623e1b675\") " pod="openstack-operators/563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv" Oct 07 13:58:46 crc kubenswrapper[4959]: I1007 13:58:46.669766 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcqcb\" (UniqueName: \"kubernetes.io/projected/209faffc-bcb0-4a9b-8558-eab623e1b675-kube-api-access-xcqcb\") pod \"563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv\" (UID: \"209faffc-bcb0-4a9b-8558-eab623e1b675\") " pod="openstack-operators/563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv" Oct 07 13:58:46 crc kubenswrapper[4959]: I1007 13:58:46.845224 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv" Oct 07 13:58:47 crc kubenswrapper[4959]: I1007 13:58:47.060610 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv"] Oct 07 13:58:47 crc kubenswrapper[4959]: W1007 13:58:47.064594 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod209faffc_bcb0_4a9b_8558_eab623e1b675.slice/crio-88bd5df0ba26809a4a0d0a150d6118a4d3ee3027d54a2606cee8eb6cee4736df WatchSource:0}: Error finding container 88bd5df0ba26809a4a0d0a150d6118a4d3ee3027d54a2606cee8eb6cee4736df: Status 404 returned error can't find the container with id 88bd5df0ba26809a4a0d0a150d6118a4d3ee3027d54a2606cee8eb6cee4736df Oct 07 13:58:47 crc kubenswrapper[4959]: I1007 13:58:47.988133 4959 generic.go:334] "Generic (PLEG): container finished" podID="6380bfbe-af4a-4a1e-a61b-cf57fbae3e72" containerID="92c9eb26776a5253df384a2e53419b0ba4da765a8362920676f14212aa5b3a97" exitCode=0 Oct 07 13:58:47 crc kubenswrapper[4959]: I1007 13:58:47.988531 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ljhpw" event={"ID":"6380bfbe-af4a-4a1e-a61b-cf57fbae3e72","Type":"ContainerDied","Data":"92c9eb26776a5253df384a2e53419b0ba4da765a8362920676f14212aa5b3a97"} Oct 07 13:58:47 crc kubenswrapper[4959]: I1007 13:58:47.994910 4959 generic.go:334] "Generic (PLEG): container finished" podID="209faffc-bcb0-4a9b-8558-eab623e1b675" containerID="39e5cc39f8e6459673f6a8e014686684219210e236ed60bbfc9af46447da28a1" exitCode=0 Oct 07 13:58:47 crc kubenswrapper[4959]: I1007 13:58:47.994978 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv" event={"ID":"209faffc-bcb0-4a9b-8558-eab623e1b675","Type":"ContainerDied","Data":"39e5cc39f8e6459673f6a8e014686684219210e236ed60bbfc9af46447da28a1"} Oct 07 13:58:47 crc kubenswrapper[4959]: I1007 13:58:47.995015 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv" event={"ID":"209faffc-bcb0-4a9b-8558-eab623e1b675","Type":"ContainerStarted","Data":"88bd5df0ba26809a4a0d0a150d6118a4d3ee3027d54a2606cee8eb6cee4736df"} Oct 07 13:58:50 crc kubenswrapper[4959]: I1007 13:58:50.010205 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ljhpw" event={"ID":"6380bfbe-af4a-4a1e-a61b-cf57fbae3e72","Type":"ContainerStarted","Data":"a3aa3cb540b9c2a05c581dd1e5148e2cd8dcff522709bf068c78dd9af2af3036"} Oct 07 13:58:50 crc kubenswrapper[4959]: I1007 13:58:50.053651 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ljhpw" podStartSLOduration=2.611980479 podStartE2EDuration="7.053607888s" podCreationTimestamp="2025-10-07 13:58:43 +0000 UTC" firstStartedPulling="2025-10-07 13:58:44.953754123 +0000 UTC m=+807.037158478" lastFinishedPulling="2025-10-07 13:58:49.395381562 +0000 UTC m=+811.478785887" observedRunningTime="2025-10-07 13:58:50.04965656 +0000 UTC m=+812.133060925" watchObservedRunningTime="2025-10-07 13:58:50.053607888 +0000 UTC m=+812.137012253" Oct 07 13:58:51 crc kubenswrapper[4959]: I1007 13:58:51.019425 4959 generic.go:334] "Generic (PLEG): container finished" podID="209faffc-bcb0-4a9b-8558-eab623e1b675" containerID="d063321d084a459ea97f24fd1c80b5e244a35cb4ff8731819ca1e478c32cfacf" exitCode=0 Oct 07 13:58:51 crc kubenswrapper[4959]: I1007 13:58:51.019653 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv" event={"ID":"209faffc-bcb0-4a9b-8558-eab623e1b675","Type":"ContainerDied","Data":"d063321d084a459ea97f24fd1c80b5e244a35cb4ff8731819ca1e478c32cfacf"} Oct 07 13:58:52 crc kubenswrapper[4959]: I1007 13:58:52.029955 4959 generic.go:334] "Generic (PLEG): container finished" podID="209faffc-bcb0-4a9b-8558-eab623e1b675" containerID="37d89a244cc8384764530fe8c0fd02737cb4bba2e1ab3fc6356d7b456391d52f" exitCode=0 Oct 07 13:58:52 crc kubenswrapper[4959]: I1007 13:58:52.030035 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv" event={"ID":"209faffc-bcb0-4a9b-8558-eab623e1b675","Type":"ContainerDied","Data":"37d89a244cc8384764530fe8c0fd02737cb4bba2e1ab3fc6356d7b456391d52f"} Oct 07 13:58:53 crc kubenswrapper[4959]: I1007 13:58:53.428951 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv" Oct 07 13:58:53 crc kubenswrapper[4959]: I1007 13:58:53.556607 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/209faffc-bcb0-4a9b-8558-eab623e1b675-util\") pod \"209faffc-bcb0-4a9b-8558-eab623e1b675\" (UID: \"209faffc-bcb0-4a9b-8558-eab623e1b675\") " Oct 07 13:58:53 crc kubenswrapper[4959]: I1007 13:58:53.556672 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcqcb\" (UniqueName: \"kubernetes.io/projected/209faffc-bcb0-4a9b-8558-eab623e1b675-kube-api-access-xcqcb\") pod \"209faffc-bcb0-4a9b-8558-eab623e1b675\" (UID: \"209faffc-bcb0-4a9b-8558-eab623e1b675\") " Oct 07 13:58:53 crc kubenswrapper[4959]: I1007 13:58:53.556803 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/209faffc-bcb0-4a9b-8558-eab623e1b675-bundle\") pod \"209faffc-bcb0-4a9b-8558-eab623e1b675\" (UID: \"209faffc-bcb0-4a9b-8558-eab623e1b675\") " Oct 07 13:58:53 crc kubenswrapper[4959]: I1007 13:58:53.557741 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/209faffc-bcb0-4a9b-8558-eab623e1b675-bundle" (OuterVolumeSpecName: "bundle") pod "209faffc-bcb0-4a9b-8558-eab623e1b675" (UID: "209faffc-bcb0-4a9b-8558-eab623e1b675"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:58:53 crc kubenswrapper[4959]: I1007 13:58:53.566021 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/209faffc-bcb0-4a9b-8558-eab623e1b675-kube-api-access-xcqcb" (OuterVolumeSpecName: "kube-api-access-xcqcb") pod "209faffc-bcb0-4a9b-8558-eab623e1b675" (UID: "209faffc-bcb0-4a9b-8558-eab623e1b675"). InnerVolumeSpecName "kube-api-access-xcqcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:58:53 crc kubenswrapper[4959]: I1007 13:58:53.567980 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/209faffc-bcb0-4a9b-8558-eab623e1b675-util" (OuterVolumeSpecName: "util") pod "209faffc-bcb0-4a9b-8558-eab623e1b675" (UID: "209faffc-bcb0-4a9b-8558-eab623e1b675"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:58:53 crc kubenswrapper[4959]: I1007 13:58:53.658482 4959 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/209faffc-bcb0-4a9b-8558-eab623e1b675-util\") on node \"crc\" DevicePath \"\"" Oct 07 13:58:53 crc kubenswrapper[4959]: I1007 13:58:53.658561 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcqcb\" (UniqueName: \"kubernetes.io/projected/209faffc-bcb0-4a9b-8558-eab623e1b675-kube-api-access-xcqcb\") on node \"crc\" DevicePath \"\"" Oct 07 13:58:53 crc kubenswrapper[4959]: I1007 13:58:53.658598 4959 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/209faffc-bcb0-4a9b-8558-eab623e1b675-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 13:58:53 crc kubenswrapper[4959]: I1007 13:58:53.825920 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ljhpw" Oct 07 13:58:53 crc kubenswrapper[4959]: I1007 13:58:53.827231 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ljhpw" Oct 07 13:58:53 crc kubenswrapper[4959]: I1007 13:58:53.872624 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ljhpw" Oct 07 13:58:54 crc kubenswrapper[4959]: I1007 13:58:54.050771 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv" event={"ID":"209faffc-bcb0-4a9b-8558-eab623e1b675","Type":"ContainerDied","Data":"88bd5df0ba26809a4a0d0a150d6118a4d3ee3027d54a2606cee8eb6cee4736df"} Oct 07 13:58:54 crc kubenswrapper[4959]: I1007 13:58:54.050849 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88bd5df0ba26809a4a0d0a150d6118a4d3ee3027d54a2606cee8eb6cee4736df" Oct 07 13:58:54 crc kubenswrapper[4959]: I1007 13:58:54.050811 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv" Oct 07 13:58:54 crc kubenswrapper[4959]: I1007 13:58:54.107287 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ljhpw" Oct 07 13:58:55 crc kubenswrapper[4959]: I1007 13:58:55.277390 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ljhpw"] Oct 07 13:58:56 crc kubenswrapper[4959]: I1007 13:58:56.064731 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ljhpw" podUID="6380bfbe-af4a-4a1e-a61b-cf57fbae3e72" containerName="registry-server" containerID="cri-o://a3aa3cb540b9c2a05c581dd1e5148e2cd8dcff522709bf068c78dd9af2af3036" gracePeriod=2 Oct 07 13:58:56 crc kubenswrapper[4959]: I1007 13:58:56.468481 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ljhpw" Oct 07 13:58:56 crc kubenswrapper[4959]: I1007 13:58:56.621249 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6380bfbe-af4a-4a1e-a61b-cf57fbae3e72-catalog-content\") pod \"6380bfbe-af4a-4a1e-a61b-cf57fbae3e72\" (UID: \"6380bfbe-af4a-4a1e-a61b-cf57fbae3e72\") " Oct 07 13:58:56 crc kubenswrapper[4959]: I1007 13:58:56.621509 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6380bfbe-af4a-4a1e-a61b-cf57fbae3e72-utilities\") pod \"6380bfbe-af4a-4a1e-a61b-cf57fbae3e72\" (UID: \"6380bfbe-af4a-4a1e-a61b-cf57fbae3e72\") " Oct 07 13:58:56 crc kubenswrapper[4959]: I1007 13:58:56.621591 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjffb\" (UniqueName: \"kubernetes.io/projected/6380bfbe-af4a-4a1e-a61b-cf57fbae3e72-kube-api-access-xjffb\") pod \"6380bfbe-af4a-4a1e-a61b-cf57fbae3e72\" (UID: \"6380bfbe-af4a-4a1e-a61b-cf57fbae3e72\") " Oct 07 13:58:56 crc kubenswrapper[4959]: I1007 13:58:56.623726 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6380bfbe-af4a-4a1e-a61b-cf57fbae3e72-utilities" (OuterVolumeSpecName: "utilities") pod "6380bfbe-af4a-4a1e-a61b-cf57fbae3e72" (UID: "6380bfbe-af4a-4a1e-a61b-cf57fbae3e72"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:58:56 crc kubenswrapper[4959]: I1007 13:58:56.647333 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6380bfbe-af4a-4a1e-a61b-cf57fbae3e72-kube-api-access-xjffb" (OuterVolumeSpecName: "kube-api-access-xjffb") pod "6380bfbe-af4a-4a1e-a61b-cf57fbae3e72" (UID: "6380bfbe-af4a-4a1e-a61b-cf57fbae3e72"). InnerVolumeSpecName "kube-api-access-xjffb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:58:56 crc kubenswrapper[4959]: I1007 13:58:56.689219 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6380bfbe-af4a-4a1e-a61b-cf57fbae3e72-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6380bfbe-af4a-4a1e-a61b-cf57fbae3e72" (UID: "6380bfbe-af4a-4a1e-a61b-cf57fbae3e72"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:58:56 crc kubenswrapper[4959]: I1007 13:58:56.722854 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6380bfbe-af4a-4a1e-a61b-cf57fbae3e72-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 13:58:56 crc kubenswrapper[4959]: I1007 13:58:56.722896 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjffb\" (UniqueName: \"kubernetes.io/projected/6380bfbe-af4a-4a1e-a61b-cf57fbae3e72-kube-api-access-xjffb\") on node \"crc\" DevicePath \"\"" Oct 07 13:58:56 crc kubenswrapper[4959]: I1007 13:58:56.722908 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6380bfbe-af4a-4a1e-a61b-cf57fbae3e72-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 13:58:57 crc kubenswrapper[4959]: I1007 13:58:57.075176 4959 generic.go:334] "Generic (PLEG): container finished" podID="6380bfbe-af4a-4a1e-a61b-cf57fbae3e72" containerID="a3aa3cb540b9c2a05c581dd1e5148e2cd8dcff522709bf068c78dd9af2af3036" exitCode=0 Oct 07 13:58:57 crc kubenswrapper[4959]: I1007 13:58:57.075243 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ljhpw" event={"ID":"6380bfbe-af4a-4a1e-a61b-cf57fbae3e72","Type":"ContainerDied","Data":"a3aa3cb540b9c2a05c581dd1e5148e2cd8dcff522709bf068c78dd9af2af3036"} Oct 07 13:58:57 crc kubenswrapper[4959]: I1007 13:58:57.075275 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ljhpw" event={"ID":"6380bfbe-af4a-4a1e-a61b-cf57fbae3e72","Type":"ContainerDied","Data":"c410e74664e9a36579ea5f0c70e87b6083a97e34a23d5d494ca9da7abf6a6e32"} Oct 07 13:58:57 crc kubenswrapper[4959]: I1007 13:58:57.075297 4959 scope.go:117] "RemoveContainer" containerID="a3aa3cb540b9c2a05c581dd1e5148e2cd8dcff522709bf068c78dd9af2af3036" Oct 07 13:58:57 crc kubenswrapper[4959]: I1007 13:58:57.075454 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ljhpw" Oct 07 13:58:57 crc kubenswrapper[4959]: I1007 13:58:57.096455 4959 scope.go:117] "RemoveContainer" containerID="92c9eb26776a5253df384a2e53419b0ba4da765a8362920676f14212aa5b3a97" Oct 07 13:58:57 crc kubenswrapper[4959]: I1007 13:58:57.107899 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ljhpw"] Oct 07 13:58:57 crc kubenswrapper[4959]: I1007 13:58:57.112658 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ljhpw"] Oct 07 13:58:57 crc kubenswrapper[4959]: I1007 13:58:57.121288 4959 scope.go:117] "RemoveContainer" containerID="c0e7dbeac848a5694353af199602ae125a49d3cd4f10e63aa814ed5306a05262" Oct 07 13:58:57 crc kubenswrapper[4959]: I1007 13:58:57.136846 4959 scope.go:117] "RemoveContainer" containerID="a3aa3cb540b9c2a05c581dd1e5148e2cd8dcff522709bf068c78dd9af2af3036" Oct 07 13:58:57 crc kubenswrapper[4959]: E1007 13:58:57.137167 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3aa3cb540b9c2a05c581dd1e5148e2cd8dcff522709bf068c78dd9af2af3036\": container with ID starting with a3aa3cb540b9c2a05c581dd1e5148e2cd8dcff522709bf068c78dd9af2af3036 not found: ID does not exist" containerID="a3aa3cb540b9c2a05c581dd1e5148e2cd8dcff522709bf068c78dd9af2af3036" Oct 07 13:58:57 crc kubenswrapper[4959]: I1007 13:58:57.137198 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3aa3cb540b9c2a05c581dd1e5148e2cd8dcff522709bf068c78dd9af2af3036"} err="failed to get container status \"a3aa3cb540b9c2a05c581dd1e5148e2cd8dcff522709bf068c78dd9af2af3036\": rpc error: code = NotFound desc = could not find container \"a3aa3cb540b9c2a05c581dd1e5148e2cd8dcff522709bf068c78dd9af2af3036\": container with ID starting with a3aa3cb540b9c2a05c581dd1e5148e2cd8dcff522709bf068c78dd9af2af3036 not found: ID does not exist" Oct 07 13:58:57 crc kubenswrapper[4959]: I1007 13:58:57.137218 4959 scope.go:117] "RemoveContainer" containerID="92c9eb26776a5253df384a2e53419b0ba4da765a8362920676f14212aa5b3a97" Oct 07 13:58:57 crc kubenswrapper[4959]: E1007 13:58:57.137833 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92c9eb26776a5253df384a2e53419b0ba4da765a8362920676f14212aa5b3a97\": container with ID starting with 92c9eb26776a5253df384a2e53419b0ba4da765a8362920676f14212aa5b3a97 not found: ID does not exist" containerID="92c9eb26776a5253df384a2e53419b0ba4da765a8362920676f14212aa5b3a97" Oct 07 13:58:57 crc kubenswrapper[4959]: I1007 13:58:57.137851 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92c9eb26776a5253df384a2e53419b0ba4da765a8362920676f14212aa5b3a97"} err="failed to get container status \"92c9eb26776a5253df384a2e53419b0ba4da765a8362920676f14212aa5b3a97\": rpc error: code = NotFound desc = could not find container \"92c9eb26776a5253df384a2e53419b0ba4da765a8362920676f14212aa5b3a97\": container with ID starting with 92c9eb26776a5253df384a2e53419b0ba4da765a8362920676f14212aa5b3a97 not found: ID does not exist" Oct 07 13:58:57 crc kubenswrapper[4959]: I1007 13:58:57.137863 4959 scope.go:117] "RemoveContainer" containerID="c0e7dbeac848a5694353af199602ae125a49d3cd4f10e63aa814ed5306a05262" Oct 07 13:58:57 crc kubenswrapper[4959]: E1007 13:58:57.138349 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0e7dbeac848a5694353af199602ae125a49d3cd4f10e63aa814ed5306a05262\": container with ID starting with c0e7dbeac848a5694353af199602ae125a49d3cd4f10e63aa814ed5306a05262 not found: ID does not exist" containerID="c0e7dbeac848a5694353af199602ae125a49d3cd4f10e63aa814ed5306a05262" Oct 07 13:58:57 crc kubenswrapper[4959]: I1007 13:58:57.138416 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0e7dbeac848a5694353af199602ae125a49d3cd4f10e63aa814ed5306a05262"} err="failed to get container status \"c0e7dbeac848a5694353af199602ae125a49d3cd4f10e63aa814ed5306a05262\": rpc error: code = NotFound desc = could not find container \"c0e7dbeac848a5694353af199602ae125a49d3cd4f10e63aa814ed5306a05262\": container with ID starting with c0e7dbeac848a5694353af199602ae125a49d3cd4f10e63aa814ed5306a05262 not found: ID does not exist" Oct 07 13:58:57 crc kubenswrapper[4959]: I1007 13:58:57.845589 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-57bc4467bb-cdctp"] Oct 07 13:58:57 crc kubenswrapper[4959]: E1007 13:58:57.845847 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6380bfbe-af4a-4a1e-a61b-cf57fbae3e72" containerName="extract-content" Oct 07 13:58:57 crc kubenswrapper[4959]: I1007 13:58:57.845859 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="6380bfbe-af4a-4a1e-a61b-cf57fbae3e72" containerName="extract-content" Oct 07 13:58:57 crc kubenswrapper[4959]: E1007 13:58:57.845873 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="209faffc-bcb0-4a9b-8558-eab623e1b675" containerName="util" Oct 07 13:58:57 crc kubenswrapper[4959]: I1007 13:58:57.845880 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="209faffc-bcb0-4a9b-8558-eab623e1b675" containerName="util" Oct 07 13:58:57 crc kubenswrapper[4959]: E1007 13:58:57.845890 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6380bfbe-af4a-4a1e-a61b-cf57fbae3e72" containerName="registry-server" Oct 07 13:58:57 crc kubenswrapper[4959]: I1007 13:58:57.845898 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="6380bfbe-af4a-4a1e-a61b-cf57fbae3e72" containerName="registry-server" Oct 07 13:58:57 crc kubenswrapper[4959]: E1007 13:58:57.845914 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="209faffc-bcb0-4a9b-8558-eab623e1b675" containerName="pull" Oct 07 13:58:57 crc kubenswrapper[4959]: I1007 13:58:57.845921 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="209faffc-bcb0-4a9b-8558-eab623e1b675" containerName="pull" Oct 07 13:58:57 crc kubenswrapper[4959]: E1007 13:58:57.845932 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="209faffc-bcb0-4a9b-8558-eab623e1b675" containerName="extract" Oct 07 13:58:57 crc kubenswrapper[4959]: I1007 13:58:57.845939 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="209faffc-bcb0-4a9b-8558-eab623e1b675" containerName="extract" Oct 07 13:58:57 crc kubenswrapper[4959]: E1007 13:58:57.845960 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6380bfbe-af4a-4a1e-a61b-cf57fbae3e72" containerName="extract-utilities" Oct 07 13:58:57 crc kubenswrapper[4959]: I1007 13:58:57.845968 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="6380bfbe-af4a-4a1e-a61b-cf57fbae3e72" containerName="extract-utilities" Oct 07 13:58:57 crc kubenswrapper[4959]: I1007 13:58:57.846120 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="6380bfbe-af4a-4a1e-a61b-cf57fbae3e72" containerName="registry-server" Oct 07 13:58:57 crc kubenswrapper[4959]: I1007 13:58:57.846142 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="209faffc-bcb0-4a9b-8558-eab623e1b675" containerName="extract" Oct 07 13:58:57 crc kubenswrapper[4959]: I1007 13:58:57.846840 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-57bc4467bb-cdctp" Oct 07 13:58:57 crc kubenswrapper[4959]: I1007 13:58:57.848638 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-2h479" Oct 07 13:58:57 crc kubenswrapper[4959]: I1007 13:58:57.869001 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-57bc4467bb-cdctp"] Oct 07 13:58:57 crc kubenswrapper[4959]: I1007 13:58:57.936701 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwlqc\" (UniqueName: \"kubernetes.io/projected/34a5b41b-5e32-42e7-a00a-9942fd069d2b-kube-api-access-dwlqc\") pod \"openstack-operator-controller-operator-57bc4467bb-cdctp\" (UID: \"34a5b41b-5e32-42e7-a00a-9942fd069d2b\") " pod="openstack-operators/openstack-operator-controller-operator-57bc4467bb-cdctp" Oct 07 13:58:58 crc kubenswrapper[4959]: I1007 13:58:58.038033 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwlqc\" (UniqueName: \"kubernetes.io/projected/34a5b41b-5e32-42e7-a00a-9942fd069d2b-kube-api-access-dwlqc\") pod \"openstack-operator-controller-operator-57bc4467bb-cdctp\" (UID: \"34a5b41b-5e32-42e7-a00a-9942fd069d2b\") " pod="openstack-operators/openstack-operator-controller-operator-57bc4467bb-cdctp" Oct 07 13:58:58 crc kubenswrapper[4959]: I1007 13:58:58.057161 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwlqc\" (UniqueName: \"kubernetes.io/projected/34a5b41b-5e32-42e7-a00a-9942fd069d2b-kube-api-access-dwlqc\") pod \"openstack-operator-controller-operator-57bc4467bb-cdctp\" (UID: \"34a5b41b-5e32-42e7-a00a-9942fd069d2b\") " pod="openstack-operators/openstack-operator-controller-operator-57bc4467bb-cdctp" Oct 07 13:58:58 crc kubenswrapper[4959]: I1007 13:58:58.163015 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-57bc4467bb-cdctp" Oct 07 13:58:58 crc kubenswrapper[4959]: I1007 13:58:58.570491 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-57bc4467bb-cdctp"] Oct 07 13:58:58 crc kubenswrapper[4959]: I1007 13:58:58.660785 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6380bfbe-af4a-4a1e-a61b-cf57fbae3e72" path="/var/lib/kubelet/pods/6380bfbe-af4a-4a1e-a61b-cf57fbae3e72/volumes" Oct 07 13:58:59 crc kubenswrapper[4959]: I1007 13:58:59.091510 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-57bc4467bb-cdctp" event={"ID":"34a5b41b-5e32-42e7-a00a-9942fd069d2b","Type":"ContainerStarted","Data":"cfce20a2959e727db7ec7895039c502645d4eb6e64bb532815dae078de4a448a"} Oct 07 13:59:03 crc kubenswrapper[4959]: I1007 13:59:03.122581 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-57bc4467bb-cdctp" event={"ID":"34a5b41b-5e32-42e7-a00a-9942fd069d2b","Type":"ContainerStarted","Data":"3eadc7632c0f224744bee6c3e076dcd53e1d7cbfa2f4c07b399d1d4c8a4cdada"} Oct 07 13:59:05 crc kubenswrapper[4959]: I1007 13:59:05.686563 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fhnzd"] Oct 07 13:59:05 crc kubenswrapper[4959]: I1007 13:59:05.689885 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fhnzd" Oct 07 13:59:05 crc kubenswrapper[4959]: I1007 13:59:05.697088 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fhnzd"] Oct 07 13:59:05 crc kubenswrapper[4959]: I1007 13:59:05.810267 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nb54l\" (UniqueName: \"kubernetes.io/projected/785aa65b-d171-4fbc-91d8-ff8288626ad6-kube-api-access-nb54l\") pod \"redhat-marketplace-fhnzd\" (UID: \"785aa65b-d171-4fbc-91d8-ff8288626ad6\") " pod="openshift-marketplace/redhat-marketplace-fhnzd" Oct 07 13:59:05 crc kubenswrapper[4959]: I1007 13:59:05.810647 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/785aa65b-d171-4fbc-91d8-ff8288626ad6-utilities\") pod \"redhat-marketplace-fhnzd\" (UID: \"785aa65b-d171-4fbc-91d8-ff8288626ad6\") " pod="openshift-marketplace/redhat-marketplace-fhnzd" Oct 07 13:59:05 crc kubenswrapper[4959]: I1007 13:59:05.810719 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/785aa65b-d171-4fbc-91d8-ff8288626ad6-catalog-content\") pod \"redhat-marketplace-fhnzd\" (UID: \"785aa65b-d171-4fbc-91d8-ff8288626ad6\") " pod="openshift-marketplace/redhat-marketplace-fhnzd" Oct 07 13:59:05 crc kubenswrapper[4959]: I1007 13:59:05.911731 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/785aa65b-d171-4fbc-91d8-ff8288626ad6-utilities\") pod \"redhat-marketplace-fhnzd\" (UID: \"785aa65b-d171-4fbc-91d8-ff8288626ad6\") " pod="openshift-marketplace/redhat-marketplace-fhnzd" Oct 07 13:59:05 crc kubenswrapper[4959]: I1007 13:59:05.911784 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/785aa65b-d171-4fbc-91d8-ff8288626ad6-catalog-content\") pod \"redhat-marketplace-fhnzd\" (UID: \"785aa65b-d171-4fbc-91d8-ff8288626ad6\") " pod="openshift-marketplace/redhat-marketplace-fhnzd" Oct 07 13:59:05 crc kubenswrapper[4959]: I1007 13:59:05.911836 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nb54l\" (UniqueName: \"kubernetes.io/projected/785aa65b-d171-4fbc-91d8-ff8288626ad6-kube-api-access-nb54l\") pod \"redhat-marketplace-fhnzd\" (UID: \"785aa65b-d171-4fbc-91d8-ff8288626ad6\") " pod="openshift-marketplace/redhat-marketplace-fhnzd" Oct 07 13:59:05 crc kubenswrapper[4959]: I1007 13:59:05.913369 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/785aa65b-d171-4fbc-91d8-ff8288626ad6-utilities\") pod \"redhat-marketplace-fhnzd\" (UID: \"785aa65b-d171-4fbc-91d8-ff8288626ad6\") " pod="openshift-marketplace/redhat-marketplace-fhnzd" Oct 07 13:59:05 crc kubenswrapper[4959]: I1007 13:59:05.913490 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/785aa65b-d171-4fbc-91d8-ff8288626ad6-catalog-content\") pod \"redhat-marketplace-fhnzd\" (UID: \"785aa65b-d171-4fbc-91d8-ff8288626ad6\") " pod="openshift-marketplace/redhat-marketplace-fhnzd" Oct 07 13:59:05 crc kubenswrapper[4959]: I1007 13:59:05.935310 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nb54l\" (UniqueName: \"kubernetes.io/projected/785aa65b-d171-4fbc-91d8-ff8288626ad6-kube-api-access-nb54l\") pod \"redhat-marketplace-fhnzd\" (UID: \"785aa65b-d171-4fbc-91d8-ff8288626ad6\") " pod="openshift-marketplace/redhat-marketplace-fhnzd" Oct 07 13:59:06 crc kubenswrapper[4959]: I1007 13:59:06.011864 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fhnzd" Oct 07 13:59:06 crc kubenswrapper[4959]: I1007 13:59:06.145520 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-57bc4467bb-cdctp" event={"ID":"34a5b41b-5e32-42e7-a00a-9942fd069d2b","Type":"ContainerStarted","Data":"5918eddd4d8d7f881902acc813928866b3fa0687a6f0763098b27756f1317ad8"} Oct 07 13:59:06 crc kubenswrapper[4959]: I1007 13:59:06.145891 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-57bc4467bb-cdctp" Oct 07 13:59:06 crc kubenswrapper[4959]: I1007 13:59:06.179816 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-57bc4467bb-cdctp" podStartSLOduration=2.604983747 podStartE2EDuration="9.179795425s" podCreationTimestamp="2025-10-07 13:58:57 +0000 UTC" firstStartedPulling="2025-10-07 13:58:58.581168475 +0000 UTC m=+820.664572800" lastFinishedPulling="2025-10-07 13:59:05.155980133 +0000 UTC m=+827.239384478" observedRunningTime="2025-10-07 13:59:06.174012843 +0000 UTC m=+828.257417168" watchObservedRunningTime="2025-10-07 13:59:06.179795425 +0000 UTC m=+828.263199750" Oct 07 13:59:06 crc kubenswrapper[4959]: I1007 13:59:06.272835 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fhnzd"] Oct 07 13:59:06 crc kubenswrapper[4959]: W1007 13:59:06.282065 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod785aa65b_d171_4fbc_91d8_ff8288626ad6.slice/crio-5be3637478c1588cbd8b0024d0813f1e7cb7cb843bc757995253c27dead6ea8f WatchSource:0}: Error finding container 5be3637478c1588cbd8b0024d0813f1e7cb7cb843bc757995253c27dead6ea8f: Status 404 returned error can't find the container with id 5be3637478c1588cbd8b0024d0813f1e7cb7cb843bc757995253c27dead6ea8f Oct 07 13:59:07 crc kubenswrapper[4959]: I1007 13:59:07.157856 4959 generic.go:334] "Generic (PLEG): container finished" podID="785aa65b-d171-4fbc-91d8-ff8288626ad6" containerID="a92eed10cfbf303b69f04fc623cca98956f0cdef64ba7ee2b4fe10e16f3130ac" exitCode=0 Oct 07 13:59:07 crc kubenswrapper[4959]: I1007 13:59:07.158160 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fhnzd" event={"ID":"785aa65b-d171-4fbc-91d8-ff8288626ad6","Type":"ContainerDied","Data":"a92eed10cfbf303b69f04fc623cca98956f0cdef64ba7ee2b4fe10e16f3130ac"} Oct 07 13:59:07 crc kubenswrapper[4959]: I1007 13:59:07.158536 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fhnzd" event={"ID":"785aa65b-d171-4fbc-91d8-ff8288626ad6","Type":"ContainerStarted","Data":"5be3637478c1588cbd8b0024d0813f1e7cb7cb843bc757995253c27dead6ea8f"} Oct 07 13:59:07 crc kubenswrapper[4959]: I1007 13:59:07.161029 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-57bc4467bb-cdctp" Oct 07 13:59:08 crc kubenswrapper[4959]: I1007 13:59:08.165961 4959 generic.go:334] "Generic (PLEG): container finished" podID="785aa65b-d171-4fbc-91d8-ff8288626ad6" containerID="28a09f24860d4ca2a0353b1ddd2d7b68dfcc147d93d388a26b8647b1db09841a" exitCode=0 Oct 07 13:59:08 crc kubenswrapper[4959]: I1007 13:59:08.166001 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fhnzd" event={"ID":"785aa65b-d171-4fbc-91d8-ff8288626ad6","Type":"ContainerDied","Data":"28a09f24860d4ca2a0353b1ddd2d7b68dfcc147d93d388a26b8647b1db09841a"} Oct 07 13:59:08 crc kubenswrapper[4959]: I1007 13:59:08.278598 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-k7pf8"] Oct 07 13:59:08 crc kubenswrapper[4959]: I1007 13:59:08.279916 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k7pf8" Oct 07 13:59:08 crc kubenswrapper[4959]: I1007 13:59:08.328469 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k7pf8"] Oct 07 13:59:08 crc kubenswrapper[4959]: I1007 13:59:08.345727 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a34edfc4-f098-4425-8ddd-09b831689d0e-utilities\") pod \"redhat-operators-k7pf8\" (UID: \"a34edfc4-f098-4425-8ddd-09b831689d0e\") " pod="openshift-marketplace/redhat-operators-k7pf8" Oct 07 13:59:08 crc kubenswrapper[4959]: I1007 13:59:08.345987 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c5dg\" (UniqueName: \"kubernetes.io/projected/a34edfc4-f098-4425-8ddd-09b831689d0e-kube-api-access-8c5dg\") pod \"redhat-operators-k7pf8\" (UID: \"a34edfc4-f098-4425-8ddd-09b831689d0e\") " pod="openshift-marketplace/redhat-operators-k7pf8" Oct 07 13:59:08 crc kubenswrapper[4959]: I1007 13:59:08.346069 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a34edfc4-f098-4425-8ddd-09b831689d0e-catalog-content\") pod \"redhat-operators-k7pf8\" (UID: \"a34edfc4-f098-4425-8ddd-09b831689d0e\") " pod="openshift-marketplace/redhat-operators-k7pf8" Oct 07 13:59:08 crc kubenswrapper[4959]: I1007 13:59:08.447836 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a34edfc4-f098-4425-8ddd-09b831689d0e-utilities\") pod \"redhat-operators-k7pf8\" (UID: \"a34edfc4-f098-4425-8ddd-09b831689d0e\") " pod="openshift-marketplace/redhat-operators-k7pf8" Oct 07 13:59:08 crc kubenswrapper[4959]: I1007 13:59:08.447896 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c5dg\" (UniqueName: \"kubernetes.io/projected/a34edfc4-f098-4425-8ddd-09b831689d0e-kube-api-access-8c5dg\") pod \"redhat-operators-k7pf8\" (UID: \"a34edfc4-f098-4425-8ddd-09b831689d0e\") " pod="openshift-marketplace/redhat-operators-k7pf8" Oct 07 13:59:08 crc kubenswrapper[4959]: I1007 13:59:08.447923 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a34edfc4-f098-4425-8ddd-09b831689d0e-catalog-content\") pod \"redhat-operators-k7pf8\" (UID: \"a34edfc4-f098-4425-8ddd-09b831689d0e\") " pod="openshift-marketplace/redhat-operators-k7pf8" Oct 07 13:59:08 crc kubenswrapper[4959]: I1007 13:59:08.448437 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a34edfc4-f098-4425-8ddd-09b831689d0e-utilities\") pod \"redhat-operators-k7pf8\" (UID: \"a34edfc4-f098-4425-8ddd-09b831689d0e\") " pod="openshift-marketplace/redhat-operators-k7pf8" Oct 07 13:59:08 crc kubenswrapper[4959]: I1007 13:59:08.448521 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a34edfc4-f098-4425-8ddd-09b831689d0e-catalog-content\") pod \"redhat-operators-k7pf8\" (UID: \"a34edfc4-f098-4425-8ddd-09b831689d0e\") " pod="openshift-marketplace/redhat-operators-k7pf8" Oct 07 13:59:08 crc kubenswrapper[4959]: I1007 13:59:08.476453 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c5dg\" (UniqueName: \"kubernetes.io/projected/a34edfc4-f098-4425-8ddd-09b831689d0e-kube-api-access-8c5dg\") pod \"redhat-operators-k7pf8\" (UID: \"a34edfc4-f098-4425-8ddd-09b831689d0e\") " pod="openshift-marketplace/redhat-operators-k7pf8" Oct 07 13:59:08 crc kubenswrapper[4959]: I1007 13:59:08.594336 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k7pf8" Oct 07 13:59:09 crc kubenswrapper[4959]: I1007 13:59:09.066218 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k7pf8"] Oct 07 13:59:09 crc kubenswrapper[4959]: W1007 13:59:09.076872 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda34edfc4_f098_4425_8ddd_09b831689d0e.slice/crio-88ea58e799210cec92d617093ebe8cdcd18fc11219976c7ac74393077abc83cf WatchSource:0}: Error finding container 88ea58e799210cec92d617093ebe8cdcd18fc11219976c7ac74393077abc83cf: Status 404 returned error can't find the container with id 88ea58e799210cec92d617093ebe8cdcd18fc11219976c7ac74393077abc83cf Oct 07 13:59:09 crc kubenswrapper[4959]: I1007 13:59:09.175685 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fhnzd" event={"ID":"785aa65b-d171-4fbc-91d8-ff8288626ad6","Type":"ContainerStarted","Data":"5fb6b9c2781f3e31c3218402a6eae22ae21bb3f1e1eb2215df7a677476b1aabc"} Oct 07 13:59:09 crc kubenswrapper[4959]: I1007 13:59:09.177603 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k7pf8" event={"ID":"a34edfc4-f098-4425-8ddd-09b831689d0e","Type":"ContainerStarted","Data":"88ea58e799210cec92d617093ebe8cdcd18fc11219976c7ac74393077abc83cf"} Oct 07 13:59:09 crc kubenswrapper[4959]: I1007 13:59:09.200394 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fhnzd" podStartSLOduration=2.737366901 podStartE2EDuration="4.200344988s" podCreationTimestamp="2025-10-07 13:59:05 +0000 UTC" firstStartedPulling="2025-10-07 13:59:07.159195294 +0000 UTC m=+829.242599649" lastFinishedPulling="2025-10-07 13:59:08.622173411 +0000 UTC m=+830.705577736" observedRunningTime="2025-10-07 13:59:09.191774777 +0000 UTC m=+831.275179092" watchObservedRunningTime="2025-10-07 13:59:09.200344988 +0000 UTC m=+831.283749313" Oct 07 13:59:09 crc kubenswrapper[4959]: E1007 13:59:09.342730 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda34edfc4_f098_4425_8ddd_09b831689d0e.slice/crio-conmon-05c6b1cb36970518a0ee653618035df84d89f4c482e67e177d8b3aa4c2d11b4a.scope\": RecentStats: unable to find data in memory cache]" Oct 07 13:59:10 crc kubenswrapper[4959]: I1007 13:59:10.186675 4959 generic.go:334] "Generic (PLEG): container finished" podID="a34edfc4-f098-4425-8ddd-09b831689d0e" containerID="05c6b1cb36970518a0ee653618035df84d89f4c482e67e177d8b3aa4c2d11b4a" exitCode=0 Oct 07 13:59:10 crc kubenswrapper[4959]: I1007 13:59:10.186805 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k7pf8" event={"ID":"a34edfc4-f098-4425-8ddd-09b831689d0e","Type":"ContainerDied","Data":"05c6b1cb36970518a0ee653618035df84d89f4c482e67e177d8b3aa4c2d11b4a"} Oct 07 13:59:11 crc kubenswrapper[4959]: I1007 13:59:11.198731 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k7pf8" event={"ID":"a34edfc4-f098-4425-8ddd-09b831689d0e","Type":"ContainerStarted","Data":"e1b94cf8e32a8073db06f3dcd77d40877afea98f87f3f4e0c8e90afd045c2501"} Oct 07 13:59:12 crc kubenswrapper[4959]: I1007 13:59:12.212415 4959 generic.go:334] "Generic (PLEG): container finished" podID="a34edfc4-f098-4425-8ddd-09b831689d0e" containerID="e1b94cf8e32a8073db06f3dcd77d40877afea98f87f3f4e0c8e90afd045c2501" exitCode=0 Oct 07 13:59:12 crc kubenswrapper[4959]: I1007 13:59:12.212523 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k7pf8" event={"ID":"a34edfc4-f098-4425-8ddd-09b831689d0e","Type":"ContainerDied","Data":"e1b94cf8e32a8073db06f3dcd77d40877afea98f87f3f4e0c8e90afd045c2501"} Oct 07 13:59:13 crc kubenswrapper[4959]: I1007 13:59:13.221755 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k7pf8" event={"ID":"a34edfc4-f098-4425-8ddd-09b831689d0e","Type":"ContainerStarted","Data":"eb0a0bac43a47e0ced08429ceebbe30f8af35253237e04a8484a1a66d5b0bda3"} Oct 07 13:59:13 crc kubenswrapper[4959]: I1007 13:59:13.240740 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-k7pf8" podStartSLOduration=2.842127665 podStartE2EDuration="5.240687333s" podCreationTimestamp="2025-10-07 13:59:08 +0000 UTC" firstStartedPulling="2025-10-07 13:59:10.18880251 +0000 UTC m=+832.272206835" lastFinishedPulling="2025-10-07 13:59:12.587362168 +0000 UTC m=+834.670766503" observedRunningTime="2025-10-07 13:59:13.237053324 +0000 UTC m=+835.320457669" watchObservedRunningTime="2025-10-07 13:59:13.240687333 +0000 UTC m=+835.324091668" Oct 07 13:59:16 crc kubenswrapper[4959]: I1007 13:59:16.012874 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fhnzd" Oct 07 13:59:16 crc kubenswrapper[4959]: I1007 13:59:16.013203 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fhnzd" Oct 07 13:59:16 crc kubenswrapper[4959]: I1007 13:59:16.074670 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fhnzd" Oct 07 13:59:16 crc kubenswrapper[4959]: I1007 13:59:16.272402 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fhnzd" Oct 07 13:59:17 crc kubenswrapper[4959]: I1007 13:59:17.883445 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sdbwr"] Oct 07 13:59:17 crc kubenswrapper[4959]: I1007 13:59:17.885152 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sdbwr" Oct 07 13:59:17 crc kubenswrapper[4959]: I1007 13:59:17.895067 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sdbwr"] Oct 07 13:59:18 crc kubenswrapper[4959]: I1007 13:59:18.050200 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9gpw\" (UniqueName: \"kubernetes.io/projected/7d9b5491-a4be-43f7-b7e8-e232406a29bc-kube-api-access-k9gpw\") pod \"certified-operators-sdbwr\" (UID: \"7d9b5491-a4be-43f7-b7e8-e232406a29bc\") " pod="openshift-marketplace/certified-operators-sdbwr" Oct 07 13:59:18 crc kubenswrapper[4959]: I1007 13:59:18.050267 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d9b5491-a4be-43f7-b7e8-e232406a29bc-catalog-content\") pod \"certified-operators-sdbwr\" (UID: \"7d9b5491-a4be-43f7-b7e8-e232406a29bc\") " pod="openshift-marketplace/certified-operators-sdbwr" Oct 07 13:59:18 crc kubenswrapper[4959]: I1007 13:59:18.050396 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d9b5491-a4be-43f7-b7e8-e232406a29bc-utilities\") pod \"certified-operators-sdbwr\" (UID: \"7d9b5491-a4be-43f7-b7e8-e232406a29bc\") " pod="openshift-marketplace/certified-operators-sdbwr" Oct 07 13:59:18 crc kubenswrapper[4959]: I1007 13:59:18.151334 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d9b5491-a4be-43f7-b7e8-e232406a29bc-catalog-content\") pod \"certified-operators-sdbwr\" (UID: \"7d9b5491-a4be-43f7-b7e8-e232406a29bc\") " pod="openshift-marketplace/certified-operators-sdbwr" Oct 07 13:59:18 crc kubenswrapper[4959]: I1007 13:59:18.151481 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d9b5491-a4be-43f7-b7e8-e232406a29bc-utilities\") pod \"certified-operators-sdbwr\" (UID: \"7d9b5491-a4be-43f7-b7e8-e232406a29bc\") " pod="openshift-marketplace/certified-operators-sdbwr" Oct 07 13:59:18 crc kubenswrapper[4959]: I1007 13:59:18.151548 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9gpw\" (UniqueName: \"kubernetes.io/projected/7d9b5491-a4be-43f7-b7e8-e232406a29bc-kube-api-access-k9gpw\") pod \"certified-operators-sdbwr\" (UID: \"7d9b5491-a4be-43f7-b7e8-e232406a29bc\") " pod="openshift-marketplace/certified-operators-sdbwr" Oct 07 13:59:18 crc kubenswrapper[4959]: I1007 13:59:18.152243 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d9b5491-a4be-43f7-b7e8-e232406a29bc-catalog-content\") pod \"certified-operators-sdbwr\" (UID: \"7d9b5491-a4be-43f7-b7e8-e232406a29bc\") " pod="openshift-marketplace/certified-operators-sdbwr" Oct 07 13:59:18 crc kubenswrapper[4959]: I1007 13:59:18.152466 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d9b5491-a4be-43f7-b7e8-e232406a29bc-utilities\") pod \"certified-operators-sdbwr\" (UID: \"7d9b5491-a4be-43f7-b7e8-e232406a29bc\") " pod="openshift-marketplace/certified-operators-sdbwr" Oct 07 13:59:18 crc kubenswrapper[4959]: I1007 13:59:18.192364 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9gpw\" (UniqueName: \"kubernetes.io/projected/7d9b5491-a4be-43f7-b7e8-e232406a29bc-kube-api-access-k9gpw\") pod \"certified-operators-sdbwr\" (UID: \"7d9b5491-a4be-43f7-b7e8-e232406a29bc\") " pod="openshift-marketplace/certified-operators-sdbwr" Oct 07 13:59:18 crc kubenswrapper[4959]: I1007 13:59:18.264205 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sdbwr" Oct 07 13:59:18 crc kubenswrapper[4959]: I1007 13:59:18.471213 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fhnzd"] Oct 07 13:59:18 crc kubenswrapper[4959]: I1007 13:59:18.471771 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fhnzd" podUID="785aa65b-d171-4fbc-91d8-ff8288626ad6" containerName="registry-server" containerID="cri-o://5fb6b9c2781f3e31c3218402a6eae22ae21bb3f1e1eb2215df7a677476b1aabc" gracePeriod=2 Oct 07 13:59:18 crc kubenswrapper[4959]: I1007 13:59:18.515638 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sdbwr"] Oct 07 13:59:18 crc kubenswrapper[4959]: W1007 13:59:18.522684 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d9b5491_a4be_43f7_b7e8_e232406a29bc.slice/crio-68f1511f7146b67eeb6a8faac5ee804d8b426298523f1ea0bb4a218fb00a5f47 WatchSource:0}: Error finding container 68f1511f7146b67eeb6a8faac5ee804d8b426298523f1ea0bb4a218fb00a5f47: Status 404 returned error can't find the container with id 68f1511f7146b67eeb6a8faac5ee804d8b426298523f1ea0bb4a218fb00a5f47 Oct 07 13:59:18 crc kubenswrapper[4959]: I1007 13:59:18.595588 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-k7pf8" Oct 07 13:59:18 crc kubenswrapper[4959]: I1007 13:59:18.595824 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-k7pf8" Oct 07 13:59:18 crc kubenswrapper[4959]: I1007 13:59:18.673482 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-k7pf8" Oct 07 13:59:18 crc kubenswrapper[4959]: I1007 13:59:18.863148 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fhnzd" Oct 07 13:59:18 crc kubenswrapper[4959]: I1007 13:59:18.871009 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/785aa65b-d171-4fbc-91d8-ff8288626ad6-utilities\") pod \"785aa65b-d171-4fbc-91d8-ff8288626ad6\" (UID: \"785aa65b-d171-4fbc-91d8-ff8288626ad6\") " Oct 07 13:59:18 crc kubenswrapper[4959]: I1007 13:59:18.871049 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/785aa65b-d171-4fbc-91d8-ff8288626ad6-catalog-content\") pod \"785aa65b-d171-4fbc-91d8-ff8288626ad6\" (UID: \"785aa65b-d171-4fbc-91d8-ff8288626ad6\") " Oct 07 13:59:18 crc kubenswrapper[4959]: I1007 13:59:18.871112 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nb54l\" (UniqueName: \"kubernetes.io/projected/785aa65b-d171-4fbc-91d8-ff8288626ad6-kube-api-access-nb54l\") pod \"785aa65b-d171-4fbc-91d8-ff8288626ad6\" (UID: \"785aa65b-d171-4fbc-91d8-ff8288626ad6\") " Oct 07 13:59:18 crc kubenswrapper[4959]: I1007 13:59:18.871925 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/785aa65b-d171-4fbc-91d8-ff8288626ad6-utilities" (OuterVolumeSpecName: "utilities") pod "785aa65b-d171-4fbc-91d8-ff8288626ad6" (UID: "785aa65b-d171-4fbc-91d8-ff8288626ad6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:59:18 crc kubenswrapper[4959]: I1007 13:59:18.876506 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/785aa65b-d171-4fbc-91d8-ff8288626ad6-kube-api-access-nb54l" (OuterVolumeSpecName: "kube-api-access-nb54l") pod "785aa65b-d171-4fbc-91d8-ff8288626ad6" (UID: "785aa65b-d171-4fbc-91d8-ff8288626ad6"). InnerVolumeSpecName "kube-api-access-nb54l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:59:18 crc kubenswrapper[4959]: I1007 13:59:18.889393 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/785aa65b-d171-4fbc-91d8-ff8288626ad6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "785aa65b-d171-4fbc-91d8-ff8288626ad6" (UID: "785aa65b-d171-4fbc-91d8-ff8288626ad6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:59:18 crc kubenswrapper[4959]: I1007 13:59:18.972051 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/785aa65b-d171-4fbc-91d8-ff8288626ad6-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 13:59:18 crc kubenswrapper[4959]: I1007 13:59:18.972077 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/785aa65b-d171-4fbc-91d8-ff8288626ad6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 13:59:18 crc kubenswrapper[4959]: I1007 13:59:18.972088 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nb54l\" (UniqueName: \"kubernetes.io/projected/785aa65b-d171-4fbc-91d8-ff8288626ad6-kube-api-access-nb54l\") on node \"crc\" DevicePath \"\"" Oct 07 13:59:19 crc kubenswrapper[4959]: I1007 13:59:19.261145 4959 generic.go:334] "Generic (PLEG): container finished" podID="7d9b5491-a4be-43f7-b7e8-e232406a29bc" containerID="00f868ebd48fd098708bf003778743e85e3e0f5bc8a04a9b0c029c674c042695" exitCode=0 Oct 07 13:59:19 crc kubenswrapper[4959]: I1007 13:59:19.261240 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sdbwr" event={"ID":"7d9b5491-a4be-43f7-b7e8-e232406a29bc","Type":"ContainerDied","Data":"00f868ebd48fd098708bf003778743e85e3e0f5bc8a04a9b0c029c674c042695"} Oct 07 13:59:19 crc kubenswrapper[4959]: I1007 13:59:19.261275 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sdbwr" event={"ID":"7d9b5491-a4be-43f7-b7e8-e232406a29bc","Type":"ContainerStarted","Data":"68f1511f7146b67eeb6a8faac5ee804d8b426298523f1ea0bb4a218fb00a5f47"} Oct 07 13:59:19 crc kubenswrapper[4959]: I1007 13:59:19.266050 4959 generic.go:334] "Generic (PLEG): container finished" podID="785aa65b-d171-4fbc-91d8-ff8288626ad6" containerID="5fb6b9c2781f3e31c3218402a6eae22ae21bb3f1e1eb2215df7a677476b1aabc" exitCode=0 Oct 07 13:59:19 crc kubenswrapper[4959]: I1007 13:59:19.266135 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fhnzd" event={"ID":"785aa65b-d171-4fbc-91d8-ff8288626ad6","Type":"ContainerDied","Data":"5fb6b9c2781f3e31c3218402a6eae22ae21bb3f1e1eb2215df7a677476b1aabc"} Oct 07 13:59:19 crc kubenswrapper[4959]: I1007 13:59:19.266166 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fhnzd" Oct 07 13:59:19 crc kubenswrapper[4959]: I1007 13:59:19.266175 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fhnzd" event={"ID":"785aa65b-d171-4fbc-91d8-ff8288626ad6","Type":"ContainerDied","Data":"5be3637478c1588cbd8b0024d0813f1e7cb7cb843bc757995253c27dead6ea8f"} Oct 07 13:59:19 crc kubenswrapper[4959]: I1007 13:59:19.266188 4959 scope.go:117] "RemoveContainer" containerID="5fb6b9c2781f3e31c3218402a6eae22ae21bb3f1e1eb2215df7a677476b1aabc" Oct 07 13:59:19 crc kubenswrapper[4959]: I1007 13:59:19.298014 4959 scope.go:117] "RemoveContainer" containerID="28a09f24860d4ca2a0353b1ddd2d7b68dfcc147d93d388a26b8647b1db09841a" Oct 07 13:59:19 crc kubenswrapper[4959]: I1007 13:59:19.315908 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fhnzd"] Oct 07 13:59:19 crc kubenswrapper[4959]: I1007 13:59:19.318984 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fhnzd"] Oct 07 13:59:19 crc kubenswrapper[4959]: I1007 13:59:19.327351 4959 scope.go:117] "RemoveContainer" containerID="a92eed10cfbf303b69f04fc623cca98956f0cdef64ba7ee2b4fe10e16f3130ac" Oct 07 13:59:19 crc kubenswrapper[4959]: I1007 13:59:19.338772 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-k7pf8" Oct 07 13:59:19 crc kubenswrapper[4959]: I1007 13:59:19.378765 4959 scope.go:117] "RemoveContainer" containerID="5fb6b9c2781f3e31c3218402a6eae22ae21bb3f1e1eb2215df7a677476b1aabc" Oct 07 13:59:19 crc kubenswrapper[4959]: E1007 13:59:19.379320 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fb6b9c2781f3e31c3218402a6eae22ae21bb3f1e1eb2215df7a677476b1aabc\": container with ID starting with 5fb6b9c2781f3e31c3218402a6eae22ae21bb3f1e1eb2215df7a677476b1aabc not found: ID does not exist" containerID="5fb6b9c2781f3e31c3218402a6eae22ae21bb3f1e1eb2215df7a677476b1aabc" Oct 07 13:59:19 crc kubenswrapper[4959]: I1007 13:59:19.379407 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fb6b9c2781f3e31c3218402a6eae22ae21bb3f1e1eb2215df7a677476b1aabc"} err="failed to get container status \"5fb6b9c2781f3e31c3218402a6eae22ae21bb3f1e1eb2215df7a677476b1aabc\": rpc error: code = NotFound desc = could not find container \"5fb6b9c2781f3e31c3218402a6eae22ae21bb3f1e1eb2215df7a677476b1aabc\": container with ID starting with 5fb6b9c2781f3e31c3218402a6eae22ae21bb3f1e1eb2215df7a677476b1aabc not found: ID does not exist" Oct 07 13:59:19 crc kubenswrapper[4959]: I1007 13:59:19.379446 4959 scope.go:117] "RemoveContainer" containerID="28a09f24860d4ca2a0353b1ddd2d7b68dfcc147d93d388a26b8647b1db09841a" Oct 07 13:59:19 crc kubenswrapper[4959]: E1007 13:59:19.379705 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28a09f24860d4ca2a0353b1ddd2d7b68dfcc147d93d388a26b8647b1db09841a\": container with ID starting with 28a09f24860d4ca2a0353b1ddd2d7b68dfcc147d93d388a26b8647b1db09841a not found: ID does not exist" containerID="28a09f24860d4ca2a0353b1ddd2d7b68dfcc147d93d388a26b8647b1db09841a" Oct 07 13:59:19 crc kubenswrapper[4959]: I1007 13:59:19.379753 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28a09f24860d4ca2a0353b1ddd2d7b68dfcc147d93d388a26b8647b1db09841a"} err="failed to get container status \"28a09f24860d4ca2a0353b1ddd2d7b68dfcc147d93d388a26b8647b1db09841a\": rpc error: code = NotFound desc = could not find container \"28a09f24860d4ca2a0353b1ddd2d7b68dfcc147d93d388a26b8647b1db09841a\": container with ID starting with 28a09f24860d4ca2a0353b1ddd2d7b68dfcc147d93d388a26b8647b1db09841a not found: ID does not exist" Oct 07 13:59:19 crc kubenswrapper[4959]: I1007 13:59:19.379799 4959 scope.go:117] "RemoveContainer" containerID="a92eed10cfbf303b69f04fc623cca98956f0cdef64ba7ee2b4fe10e16f3130ac" Oct 07 13:59:19 crc kubenswrapper[4959]: E1007 13:59:19.380023 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a92eed10cfbf303b69f04fc623cca98956f0cdef64ba7ee2b4fe10e16f3130ac\": container with ID starting with a92eed10cfbf303b69f04fc623cca98956f0cdef64ba7ee2b4fe10e16f3130ac not found: ID does not exist" containerID="a92eed10cfbf303b69f04fc623cca98956f0cdef64ba7ee2b4fe10e16f3130ac" Oct 07 13:59:19 crc kubenswrapper[4959]: I1007 13:59:19.380049 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a92eed10cfbf303b69f04fc623cca98956f0cdef64ba7ee2b4fe10e16f3130ac"} err="failed to get container status \"a92eed10cfbf303b69f04fc623cca98956f0cdef64ba7ee2b4fe10e16f3130ac\": rpc error: code = NotFound desc = could not find container \"a92eed10cfbf303b69f04fc623cca98956f0cdef64ba7ee2b4fe10e16f3130ac\": container with ID starting with a92eed10cfbf303b69f04fc623cca98956f0cdef64ba7ee2b4fe10e16f3130ac not found: ID does not exist" Oct 07 13:59:20 crc kubenswrapper[4959]: I1007 13:59:20.276820 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sdbwr" event={"ID":"7d9b5491-a4be-43f7-b7e8-e232406a29bc","Type":"ContainerStarted","Data":"83f1647b988b29e021535133c9f73c1628480173b5452d2a41c002fc5b77fe80"} Oct 07 13:59:20 crc kubenswrapper[4959]: I1007 13:59:20.662035 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="785aa65b-d171-4fbc-91d8-ff8288626ad6" path="/var/lib/kubelet/pods/785aa65b-d171-4fbc-91d8-ff8288626ad6/volumes" Oct 07 13:59:21 crc kubenswrapper[4959]: I1007 13:59:21.286223 4959 generic.go:334] "Generic (PLEG): container finished" podID="7d9b5491-a4be-43f7-b7e8-e232406a29bc" containerID="83f1647b988b29e021535133c9f73c1628480173b5452d2a41c002fc5b77fe80" exitCode=0 Oct 07 13:59:21 crc kubenswrapper[4959]: I1007 13:59:21.286301 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sdbwr" event={"ID":"7d9b5491-a4be-43f7-b7e8-e232406a29bc","Type":"ContainerDied","Data":"83f1647b988b29e021535133c9f73c1628480173b5452d2a41c002fc5b77fe80"} Oct 07 13:59:22 crc kubenswrapper[4959]: I1007 13:59:22.270991 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k7pf8"] Oct 07 13:59:22 crc kubenswrapper[4959]: I1007 13:59:22.295615 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sdbwr" event={"ID":"7d9b5491-a4be-43f7-b7e8-e232406a29bc","Type":"ContainerStarted","Data":"dd9f196db46e082a0d556b100c414f56bd676581f3bb21e975b190bcd8dcc813"} Oct 07 13:59:22 crc kubenswrapper[4959]: I1007 13:59:22.295759 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-k7pf8" podUID="a34edfc4-f098-4425-8ddd-09b831689d0e" containerName="registry-server" containerID="cri-o://eb0a0bac43a47e0ced08429ceebbe30f8af35253237e04a8484a1a66d5b0bda3" gracePeriod=2 Oct 07 13:59:22 crc kubenswrapper[4959]: I1007 13:59:22.317122 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sdbwr" podStartSLOduration=2.885298882 podStartE2EDuration="5.317081287s" podCreationTimestamp="2025-10-07 13:59:17 +0000 UTC" firstStartedPulling="2025-10-07 13:59:19.263179583 +0000 UTC m=+841.346583908" lastFinishedPulling="2025-10-07 13:59:21.694961988 +0000 UTC m=+843.778366313" observedRunningTime="2025-10-07 13:59:22.312591266 +0000 UTC m=+844.395995621" watchObservedRunningTime="2025-10-07 13:59:22.317081287 +0000 UTC m=+844.400485642" Oct 07 13:59:24 crc kubenswrapper[4959]: I1007 13:59:24.322172 4959 generic.go:334] "Generic (PLEG): container finished" podID="a34edfc4-f098-4425-8ddd-09b831689d0e" containerID="eb0a0bac43a47e0ced08429ceebbe30f8af35253237e04a8484a1a66d5b0bda3" exitCode=0 Oct 07 13:59:24 crc kubenswrapper[4959]: I1007 13:59:24.322247 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k7pf8" event={"ID":"a34edfc4-f098-4425-8ddd-09b831689d0e","Type":"ContainerDied","Data":"eb0a0bac43a47e0ced08429ceebbe30f8af35253237e04a8484a1a66d5b0bda3"} Oct 07 13:59:24 crc kubenswrapper[4959]: I1007 13:59:24.530680 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k7pf8" Oct 07 13:59:24 crc kubenswrapper[4959]: I1007 13:59:24.555187 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8c5dg\" (UniqueName: \"kubernetes.io/projected/a34edfc4-f098-4425-8ddd-09b831689d0e-kube-api-access-8c5dg\") pod \"a34edfc4-f098-4425-8ddd-09b831689d0e\" (UID: \"a34edfc4-f098-4425-8ddd-09b831689d0e\") " Oct 07 13:59:24 crc kubenswrapper[4959]: I1007 13:59:24.555258 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a34edfc4-f098-4425-8ddd-09b831689d0e-catalog-content\") pod \"a34edfc4-f098-4425-8ddd-09b831689d0e\" (UID: \"a34edfc4-f098-4425-8ddd-09b831689d0e\") " Oct 07 13:59:24 crc kubenswrapper[4959]: I1007 13:59:24.555288 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a34edfc4-f098-4425-8ddd-09b831689d0e-utilities\") pod \"a34edfc4-f098-4425-8ddd-09b831689d0e\" (UID: \"a34edfc4-f098-4425-8ddd-09b831689d0e\") " Oct 07 13:59:24 crc kubenswrapper[4959]: I1007 13:59:24.557034 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a34edfc4-f098-4425-8ddd-09b831689d0e-utilities" (OuterVolumeSpecName: "utilities") pod "a34edfc4-f098-4425-8ddd-09b831689d0e" (UID: "a34edfc4-f098-4425-8ddd-09b831689d0e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:59:24 crc kubenswrapper[4959]: I1007 13:59:24.560711 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a34edfc4-f098-4425-8ddd-09b831689d0e-kube-api-access-8c5dg" (OuterVolumeSpecName: "kube-api-access-8c5dg") pod "a34edfc4-f098-4425-8ddd-09b831689d0e" (UID: "a34edfc4-f098-4425-8ddd-09b831689d0e"). InnerVolumeSpecName "kube-api-access-8c5dg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:59:24 crc kubenswrapper[4959]: I1007 13:59:24.657375 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8c5dg\" (UniqueName: \"kubernetes.io/projected/a34edfc4-f098-4425-8ddd-09b831689d0e-kube-api-access-8c5dg\") on node \"crc\" DevicePath \"\"" Oct 07 13:59:24 crc kubenswrapper[4959]: I1007 13:59:24.657681 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a34edfc4-f098-4425-8ddd-09b831689d0e-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 13:59:24 crc kubenswrapper[4959]: I1007 13:59:24.660969 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a34edfc4-f098-4425-8ddd-09b831689d0e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a34edfc4-f098-4425-8ddd-09b831689d0e" (UID: "a34edfc4-f098-4425-8ddd-09b831689d0e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:59:24 crc kubenswrapper[4959]: I1007 13:59:24.758884 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a34edfc4-f098-4425-8ddd-09b831689d0e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 13:59:25 crc kubenswrapper[4959]: I1007 13:59:25.333014 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k7pf8" event={"ID":"a34edfc4-f098-4425-8ddd-09b831689d0e","Type":"ContainerDied","Data":"88ea58e799210cec92d617093ebe8cdcd18fc11219976c7ac74393077abc83cf"} Oct 07 13:59:25 crc kubenswrapper[4959]: I1007 13:59:25.333096 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k7pf8" Oct 07 13:59:25 crc kubenswrapper[4959]: I1007 13:59:25.333503 4959 scope.go:117] "RemoveContainer" containerID="eb0a0bac43a47e0ced08429ceebbe30f8af35253237e04a8484a1a66d5b0bda3" Oct 07 13:59:25 crc kubenswrapper[4959]: I1007 13:59:25.354734 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k7pf8"] Oct 07 13:59:25 crc kubenswrapper[4959]: I1007 13:59:25.359602 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-k7pf8"] Oct 07 13:59:25 crc kubenswrapper[4959]: I1007 13:59:25.364335 4959 scope.go:117] "RemoveContainer" containerID="e1b94cf8e32a8073db06f3dcd77d40877afea98f87f3f4e0c8e90afd045c2501" Oct 07 13:59:25 crc kubenswrapper[4959]: I1007 13:59:25.382752 4959 scope.go:117] "RemoveContainer" containerID="05c6b1cb36970518a0ee653618035df84d89f4c482e67e177d8b3aa4c2d11b4a" Oct 07 13:59:26 crc kubenswrapper[4959]: I1007 13:59:26.662393 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a34edfc4-f098-4425-8ddd-09b831689d0e" path="/var/lib/kubelet/pods/a34edfc4-f098-4425-8ddd-09b831689d0e/volumes" Oct 07 13:59:28 crc kubenswrapper[4959]: I1007 13:59:28.265399 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sdbwr" Oct 07 13:59:28 crc kubenswrapper[4959]: I1007 13:59:28.265770 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sdbwr" Oct 07 13:59:28 crc kubenswrapper[4959]: I1007 13:59:28.314824 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sdbwr" Oct 07 13:59:28 crc kubenswrapper[4959]: I1007 13:59:28.387949 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sdbwr" Oct 07 13:59:28 crc kubenswrapper[4959]: I1007 13:59:28.875996 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sdbwr"] Oct 07 13:59:30 crc kubenswrapper[4959]: I1007 13:59:30.366429 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-sdbwr" podUID="7d9b5491-a4be-43f7-b7e8-e232406a29bc" containerName="registry-server" containerID="cri-o://dd9f196db46e082a0d556b100c414f56bd676581f3bb21e975b190bcd8dcc813" gracePeriod=2 Oct 07 13:59:31 crc kubenswrapper[4959]: I1007 13:59:31.346548 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sdbwr" Oct 07 13:59:31 crc kubenswrapper[4959]: I1007 13:59:31.373615 4959 generic.go:334] "Generic (PLEG): container finished" podID="7d9b5491-a4be-43f7-b7e8-e232406a29bc" containerID="dd9f196db46e082a0d556b100c414f56bd676581f3bb21e975b190bcd8dcc813" exitCode=0 Oct 07 13:59:31 crc kubenswrapper[4959]: I1007 13:59:31.373659 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sdbwr" event={"ID":"7d9b5491-a4be-43f7-b7e8-e232406a29bc","Type":"ContainerDied","Data":"dd9f196db46e082a0d556b100c414f56bd676581f3bb21e975b190bcd8dcc813"} Oct 07 13:59:31 crc kubenswrapper[4959]: I1007 13:59:31.373685 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sdbwr" event={"ID":"7d9b5491-a4be-43f7-b7e8-e232406a29bc","Type":"ContainerDied","Data":"68f1511f7146b67eeb6a8faac5ee804d8b426298523f1ea0bb4a218fb00a5f47"} Oct 07 13:59:31 crc kubenswrapper[4959]: I1007 13:59:31.373703 4959 scope.go:117] "RemoveContainer" containerID="dd9f196db46e082a0d556b100c414f56bd676581f3bb21e975b190bcd8dcc813" Oct 07 13:59:31 crc kubenswrapper[4959]: I1007 13:59:31.373819 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sdbwr" Oct 07 13:59:31 crc kubenswrapper[4959]: I1007 13:59:31.393486 4959 scope.go:117] "RemoveContainer" containerID="83f1647b988b29e021535133c9f73c1628480173b5452d2a41c002fc5b77fe80" Oct 07 13:59:31 crc kubenswrapper[4959]: I1007 13:59:31.407804 4959 scope.go:117] "RemoveContainer" containerID="00f868ebd48fd098708bf003778743e85e3e0f5bc8a04a9b0c029c674c042695" Oct 07 13:59:31 crc kubenswrapper[4959]: I1007 13:59:31.426166 4959 scope.go:117] "RemoveContainer" containerID="dd9f196db46e082a0d556b100c414f56bd676581f3bb21e975b190bcd8dcc813" Oct 07 13:59:31 crc kubenswrapper[4959]: E1007 13:59:31.426968 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd9f196db46e082a0d556b100c414f56bd676581f3bb21e975b190bcd8dcc813\": container with ID starting with dd9f196db46e082a0d556b100c414f56bd676581f3bb21e975b190bcd8dcc813 not found: ID does not exist" containerID="dd9f196db46e082a0d556b100c414f56bd676581f3bb21e975b190bcd8dcc813" Oct 07 13:59:31 crc kubenswrapper[4959]: I1007 13:59:31.426996 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd9f196db46e082a0d556b100c414f56bd676581f3bb21e975b190bcd8dcc813"} err="failed to get container status \"dd9f196db46e082a0d556b100c414f56bd676581f3bb21e975b190bcd8dcc813\": rpc error: code = NotFound desc = could not find container \"dd9f196db46e082a0d556b100c414f56bd676581f3bb21e975b190bcd8dcc813\": container with ID starting with dd9f196db46e082a0d556b100c414f56bd676581f3bb21e975b190bcd8dcc813 not found: ID does not exist" Oct 07 13:59:31 crc kubenswrapper[4959]: I1007 13:59:31.427017 4959 scope.go:117] "RemoveContainer" containerID="83f1647b988b29e021535133c9f73c1628480173b5452d2a41c002fc5b77fe80" Oct 07 13:59:31 crc kubenswrapper[4959]: E1007 13:59:31.427671 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83f1647b988b29e021535133c9f73c1628480173b5452d2a41c002fc5b77fe80\": container with ID starting with 83f1647b988b29e021535133c9f73c1628480173b5452d2a41c002fc5b77fe80 not found: ID does not exist" containerID="83f1647b988b29e021535133c9f73c1628480173b5452d2a41c002fc5b77fe80" Oct 07 13:59:31 crc kubenswrapper[4959]: I1007 13:59:31.427698 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83f1647b988b29e021535133c9f73c1628480173b5452d2a41c002fc5b77fe80"} err="failed to get container status \"83f1647b988b29e021535133c9f73c1628480173b5452d2a41c002fc5b77fe80\": rpc error: code = NotFound desc = could not find container \"83f1647b988b29e021535133c9f73c1628480173b5452d2a41c002fc5b77fe80\": container with ID starting with 83f1647b988b29e021535133c9f73c1628480173b5452d2a41c002fc5b77fe80 not found: ID does not exist" Oct 07 13:59:31 crc kubenswrapper[4959]: I1007 13:59:31.427714 4959 scope.go:117] "RemoveContainer" containerID="00f868ebd48fd098708bf003778743e85e3e0f5bc8a04a9b0c029c674c042695" Oct 07 13:59:31 crc kubenswrapper[4959]: E1007 13:59:31.428245 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00f868ebd48fd098708bf003778743e85e3e0f5bc8a04a9b0c029c674c042695\": container with ID starting with 00f868ebd48fd098708bf003778743e85e3e0f5bc8a04a9b0c029c674c042695 not found: ID does not exist" containerID="00f868ebd48fd098708bf003778743e85e3e0f5bc8a04a9b0c029c674c042695" Oct 07 13:59:31 crc kubenswrapper[4959]: I1007 13:59:31.428266 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00f868ebd48fd098708bf003778743e85e3e0f5bc8a04a9b0c029c674c042695"} err="failed to get container status \"00f868ebd48fd098708bf003778743e85e3e0f5bc8a04a9b0c029c674c042695\": rpc error: code = NotFound desc = could not find container \"00f868ebd48fd098708bf003778743e85e3e0f5bc8a04a9b0c029c674c042695\": container with ID starting with 00f868ebd48fd098708bf003778743e85e3e0f5bc8a04a9b0c029c674c042695 not found: ID does not exist" Oct 07 13:59:31 crc kubenswrapper[4959]: I1007 13:59:31.445901 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d9b5491-a4be-43f7-b7e8-e232406a29bc-utilities\") pod \"7d9b5491-a4be-43f7-b7e8-e232406a29bc\" (UID: \"7d9b5491-a4be-43f7-b7e8-e232406a29bc\") " Oct 07 13:59:31 crc kubenswrapper[4959]: I1007 13:59:31.446193 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9gpw\" (UniqueName: \"kubernetes.io/projected/7d9b5491-a4be-43f7-b7e8-e232406a29bc-kube-api-access-k9gpw\") pod \"7d9b5491-a4be-43f7-b7e8-e232406a29bc\" (UID: \"7d9b5491-a4be-43f7-b7e8-e232406a29bc\") " Oct 07 13:59:31 crc kubenswrapper[4959]: I1007 13:59:31.446256 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d9b5491-a4be-43f7-b7e8-e232406a29bc-catalog-content\") pod \"7d9b5491-a4be-43f7-b7e8-e232406a29bc\" (UID: \"7d9b5491-a4be-43f7-b7e8-e232406a29bc\") " Oct 07 13:59:31 crc kubenswrapper[4959]: I1007 13:59:31.446952 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d9b5491-a4be-43f7-b7e8-e232406a29bc-utilities" (OuterVolumeSpecName: "utilities") pod "7d9b5491-a4be-43f7-b7e8-e232406a29bc" (UID: "7d9b5491-a4be-43f7-b7e8-e232406a29bc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:59:31 crc kubenswrapper[4959]: I1007 13:59:31.456726 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d9b5491-a4be-43f7-b7e8-e232406a29bc-kube-api-access-k9gpw" (OuterVolumeSpecName: "kube-api-access-k9gpw") pod "7d9b5491-a4be-43f7-b7e8-e232406a29bc" (UID: "7d9b5491-a4be-43f7-b7e8-e232406a29bc"). InnerVolumeSpecName "kube-api-access-k9gpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 13:59:31 crc kubenswrapper[4959]: I1007 13:59:31.504196 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d9b5491-a4be-43f7-b7e8-e232406a29bc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7d9b5491-a4be-43f7-b7e8-e232406a29bc" (UID: "7d9b5491-a4be-43f7-b7e8-e232406a29bc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 13:59:31 crc kubenswrapper[4959]: I1007 13:59:31.548093 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d9b5491-a4be-43f7-b7e8-e232406a29bc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 13:59:31 crc kubenswrapper[4959]: I1007 13:59:31.548275 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d9b5491-a4be-43f7-b7e8-e232406a29bc-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 13:59:31 crc kubenswrapper[4959]: I1007 13:59:31.548291 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9gpw\" (UniqueName: \"kubernetes.io/projected/7d9b5491-a4be-43f7-b7e8-e232406a29bc-kube-api-access-k9gpw\") on node \"crc\" DevicePath \"\"" Oct 07 13:59:31 crc kubenswrapper[4959]: I1007 13:59:31.720739 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sdbwr"] Oct 07 13:59:31 crc kubenswrapper[4959]: I1007 13:59:31.725787 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-sdbwr"] Oct 07 13:59:32 crc kubenswrapper[4959]: I1007 13:59:32.664068 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d9b5491-a4be-43f7-b7e8-e232406a29bc" path="/var/lib/kubelet/pods/7d9b5491-a4be-43f7-b7e8-e232406a29bc/volumes" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.086008 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f56ff694-9k7x7"] Oct 07 13:59:45 crc kubenswrapper[4959]: E1007 13:59:45.086912 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="785aa65b-d171-4fbc-91d8-ff8288626ad6" containerName="registry-server" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.086924 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="785aa65b-d171-4fbc-91d8-ff8288626ad6" containerName="registry-server" Oct 07 13:59:45 crc kubenswrapper[4959]: E1007 13:59:45.086932 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="785aa65b-d171-4fbc-91d8-ff8288626ad6" containerName="extract-utilities" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.086938 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="785aa65b-d171-4fbc-91d8-ff8288626ad6" containerName="extract-utilities" Oct 07 13:59:45 crc kubenswrapper[4959]: E1007 13:59:45.086948 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a34edfc4-f098-4425-8ddd-09b831689d0e" containerName="extract-utilities" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.086956 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a34edfc4-f098-4425-8ddd-09b831689d0e" containerName="extract-utilities" Oct 07 13:59:45 crc kubenswrapper[4959]: E1007 13:59:45.086965 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a34edfc4-f098-4425-8ddd-09b831689d0e" containerName="registry-server" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.086971 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a34edfc4-f098-4425-8ddd-09b831689d0e" containerName="registry-server" Oct 07 13:59:45 crc kubenswrapper[4959]: E1007 13:59:45.086980 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a34edfc4-f098-4425-8ddd-09b831689d0e" containerName="extract-content" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.086987 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a34edfc4-f098-4425-8ddd-09b831689d0e" containerName="extract-content" Oct 07 13:59:45 crc kubenswrapper[4959]: E1007 13:59:45.086994 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d9b5491-a4be-43f7-b7e8-e232406a29bc" containerName="extract-utilities" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.087001 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d9b5491-a4be-43f7-b7e8-e232406a29bc" containerName="extract-utilities" Oct 07 13:59:45 crc kubenswrapper[4959]: E1007 13:59:45.087009 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="785aa65b-d171-4fbc-91d8-ff8288626ad6" containerName="extract-content" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.087015 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="785aa65b-d171-4fbc-91d8-ff8288626ad6" containerName="extract-content" Oct 07 13:59:45 crc kubenswrapper[4959]: E1007 13:59:45.087024 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d9b5491-a4be-43f7-b7e8-e232406a29bc" containerName="registry-server" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.087029 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d9b5491-a4be-43f7-b7e8-e232406a29bc" containerName="registry-server" Oct 07 13:59:45 crc kubenswrapper[4959]: E1007 13:59:45.087039 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d9b5491-a4be-43f7-b7e8-e232406a29bc" containerName="extract-content" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.087044 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d9b5491-a4be-43f7-b7e8-e232406a29bc" containerName="extract-content" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.087162 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a34edfc4-f098-4425-8ddd-09b831689d0e" containerName="registry-server" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.087175 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="785aa65b-d171-4fbc-91d8-ff8288626ad6" containerName="registry-server" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.087186 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d9b5491-a4be-43f7-b7e8-e232406a29bc" containerName="registry-server" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.087756 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-9k7x7" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.091258 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-xffqh" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.094670 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-84bd8f6848-85vgl"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.095518 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-85vgl" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.098147 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-xgdhc" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.114632 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f56ff694-9k7x7"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.122693 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-84bd8f6848-85vgl"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.133857 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-fd648f65-bc7rh"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.134788 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-fd648f65-bc7rh" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.134849 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhgkr\" (UniqueName: \"kubernetes.io/projected/1e3c7b8e-8a44-4539-9e01-4fe9ce17266e-kube-api-access-nhgkr\") pod \"barbican-operator-controller-manager-64f56ff694-9k7x7\" (UID: \"1e3c7b8e-8a44-4539-9e01-4fe9ce17266e\") " pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-9k7x7" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.134881 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rc72\" (UniqueName: \"kubernetes.io/projected/2b84227a-34a9-4969-97be-4750eb555f22-kube-api-access-7rc72\") pod \"cinder-operator-controller-manager-84bd8f6848-85vgl\" (UID: \"2b84227a-34a9-4969-97be-4750eb555f22\") " pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-85vgl" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.138235 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-58d86cd59d-pnbwt"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.139642 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-pnbwt" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.149189 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-7ccfc8cf49-q56bw"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.150204 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-q56bw" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.151330 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-hlh6x" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.151611 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-dr8f8" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.158033 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-cq44k" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.167418 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-fd648f65-bc7rh"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.173071 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-58d86cd59d-pnbwt"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.193551 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b477879bc-p6mgb"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.194557 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-p6mgb" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.196933 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-km9cb" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.214724 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-84788b6bc5-rdj2q"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.216027 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-7ccfc8cf49-q56bw"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.216037 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-rdj2q" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.234005 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.234243 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-bbtkk" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.236013 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-84788b6bc5-rdj2q"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.236857 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6whrp\" (UniqueName: \"kubernetes.io/projected/e46e007a-5053-443a-9619-247dd3d79a8e-kube-api-access-6whrp\") pod \"designate-operator-controller-manager-58d86cd59d-pnbwt\" (UID: \"e46e007a-5053-443a-9619-247dd3d79a8e\") " pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-pnbwt" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.236958 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhgkr\" (UniqueName: \"kubernetes.io/projected/1e3c7b8e-8a44-4539-9e01-4fe9ce17266e-kube-api-access-nhgkr\") pod \"barbican-operator-controller-manager-64f56ff694-9k7x7\" (UID: \"1e3c7b8e-8a44-4539-9e01-4fe9ce17266e\") " pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-9k7x7" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.236985 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rc72\" (UniqueName: \"kubernetes.io/projected/2b84227a-34a9-4969-97be-4750eb555f22-kube-api-access-7rc72\") pod \"cinder-operator-controller-manager-84bd8f6848-85vgl\" (UID: \"2b84227a-34a9-4969-97be-4750eb555f22\") " pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-85vgl" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.263199 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b477879bc-p6mgb"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.290549 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhgkr\" (UniqueName: \"kubernetes.io/projected/1e3c7b8e-8a44-4539-9e01-4fe9ce17266e-kube-api-access-nhgkr\") pod \"barbican-operator-controller-manager-64f56ff694-9k7x7\" (UID: \"1e3c7b8e-8a44-4539-9e01-4fe9ce17266e\") " pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-9k7x7" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.290549 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rc72\" (UniqueName: \"kubernetes.io/projected/2b84227a-34a9-4969-97be-4750eb555f22-kube-api-access-7rc72\") pod \"cinder-operator-controller-manager-84bd8f6848-85vgl\" (UID: \"2b84227a-34a9-4969-97be-4750eb555f22\") " pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-85vgl" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.295609 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5467f8988c-l4hdq"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.296587 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-l4hdq" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.302414 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-n25lb" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.305149 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5b84cc7657-k5rmd"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.306118 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-k5rmd" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.312431 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7cb48dbc-hhbqh"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.313706 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-hhbqh" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.313861 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-zqvw8" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.317940 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-tfgtw" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.323621 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5467f8988c-l4hdq"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.330932 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5b84cc7657-k5rmd"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.340880 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/77cf7bcc-63b9-4ee3-84bb-4d4b784750c6-cert\") pod \"infra-operator-controller-manager-84788b6bc5-rdj2q\" (UID: \"77cf7bcc-63b9-4ee3-84bb-4d4b784750c6\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-rdj2q" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.340919 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-264kt\" (UniqueName: \"kubernetes.io/projected/ca1a8a8b-1649-49b3-b595-0e12408673c9-kube-api-access-264kt\") pod \"glance-operator-controller-manager-fd648f65-bc7rh\" (UID: \"ca1a8a8b-1649-49b3-b595-0e12408673c9\") " pod="openstack-operators/glance-operator-controller-manager-fd648f65-bc7rh" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.340937 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbwsj\" (UniqueName: \"kubernetes.io/projected/c21b1db6-de7b-48eb-9796-b7d45a638e8c-kube-api-access-gbwsj\") pod \"horizon-operator-controller-manager-5b477879bc-p6mgb\" (UID: \"c21b1db6-de7b-48eb-9796-b7d45a638e8c\") " pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-p6mgb" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.340984 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6whrp\" (UniqueName: \"kubernetes.io/projected/e46e007a-5053-443a-9619-247dd3d79a8e-kube-api-access-6whrp\") pod \"designate-operator-controller-manager-58d86cd59d-pnbwt\" (UID: \"e46e007a-5053-443a-9619-247dd3d79a8e\") " pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-pnbwt" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.341008 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcpjt\" (UniqueName: \"kubernetes.io/projected/98825e89-a8c2-4023-b139-4c6f5d9b7987-kube-api-access-hcpjt\") pod \"heat-operator-controller-manager-7ccfc8cf49-q56bw\" (UID: \"98825e89-a8c2-4023-b139-4c6f5d9b7987\") " pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-q56bw" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.341038 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8phv\" (UniqueName: \"kubernetes.io/projected/77cf7bcc-63b9-4ee3-84bb-4d4b784750c6-kube-api-access-z8phv\") pod \"infra-operator-controller-manager-84788b6bc5-rdj2q\" (UID: \"77cf7bcc-63b9-4ee3-84bb-4d4b784750c6\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-rdj2q" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.347444 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-ksjkq"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.353129 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-ksjkq" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.354551 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-brxtd" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.357161 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-ksjkq"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.369689 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-69b956fbf6-k4pkf"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.370765 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-k4pkf" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.383174 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7cb48dbc-hhbqh"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.383250 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-69b956fbf6-k4pkf"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.387769 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6whrp\" (UniqueName: \"kubernetes.io/projected/e46e007a-5053-443a-9619-247dd3d79a8e-kube-api-access-6whrp\") pod \"designate-operator-controller-manager-58d86cd59d-pnbwt\" (UID: \"e46e007a-5053-443a-9619-247dd3d79a8e\") " pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-pnbwt" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.388514 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-xq62k" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.404076 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-6c9b57c67-mgrc2"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.405381 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-mgrc2" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.409491 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f59f9d8-tfrz7"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.410552 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-tfrz7" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.411025 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-9k7x7" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.413068 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-nqmnp" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.413549 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-lwzdf" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.417691 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-6c9b57c67-mgrc2"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.422219 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-85vgl" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.426936 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f59f9d8-tfrz7"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.431072 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.432112 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.442308 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcpjt\" (UniqueName: \"kubernetes.io/projected/98825e89-a8c2-4023-b139-4c6f5d9b7987-kube-api-access-hcpjt\") pod \"heat-operator-controller-manager-7ccfc8cf49-q56bw\" (UID: \"98825e89-a8c2-4023-b139-4c6f5d9b7987\") " pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-q56bw" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.442362 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldw5w\" (UniqueName: \"kubernetes.io/projected/c47e0f74-de21-41f6-9142-85d47fcafc60-kube-api-access-ldw5w\") pod \"ironic-operator-controller-manager-5467f8988c-l4hdq\" (UID: \"c47e0f74-de21-41f6-9142-85d47fcafc60\") " pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-l4hdq" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.442389 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8phv\" (UniqueName: \"kubernetes.io/projected/77cf7bcc-63b9-4ee3-84bb-4d4b784750c6-kube-api-access-z8phv\") pod \"infra-operator-controller-manager-84788b6bc5-rdj2q\" (UID: \"77cf7bcc-63b9-4ee3-84bb-4d4b784750c6\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-rdj2q" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.442426 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/77cf7bcc-63b9-4ee3-84bb-4d4b784750c6-cert\") pod \"infra-operator-controller-manager-84788b6bc5-rdj2q\" (UID: \"77cf7bcc-63b9-4ee3-84bb-4d4b784750c6\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-rdj2q" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.442451 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8g29\" (UniqueName: \"kubernetes.io/projected/e1da143c-64e7-4c75-97fb-4e3553fc2906-kube-api-access-n8g29\") pod \"mariadb-operator-controller-manager-d6c9dc5bc-ksjkq\" (UID: \"e1da143c-64e7-4c75-97fb-4e3553fc2906\") " pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-ksjkq" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.442470 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-264kt\" (UniqueName: \"kubernetes.io/projected/ca1a8a8b-1649-49b3-b595-0e12408673c9-kube-api-access-264kt\") pod \"glance-operator-controller-manager-fd648f65-bc7rh\" (UID: \"ca1a8a8b-1649-49b3-b595-0e12408673c9\") " pod="openstack-operators/glance-operator-controller-manager-fd648f65-bc7rh" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.442488 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbwsj\" (UniqueName: \"kubernetes.io/projected/c21b1db6-de7b-48eb-9796-b7d45a638e8c-kube-api-access-gbwsj\") pod \"horizon-operator-controller-manager-5b477879bc-p6mgb\" (UID: \"c21b1db6-de7b-48eb-9796-b7d45a638e8c\") " pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-p6mgb" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.442520 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dzkl\" (UniqueName: \"kubernetes.io/projected/801942fb-8de2-449b-9a14-4616bdec95aa-kube-api-access-5dzkl\") pod \"manila-operator-controller-manager-7cb48dbc-hhbqh\" (UID: \"801942fb-8de2-449b-9a14-4616bdec95aa\") " pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-hhbqh" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.442552 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdh85\" (UniqueName: \"kubernetes.io/projected/64565126-cba9-4da4-8e0e-7a75f2acc4f1-kube-api-access-vdh85\") pod \"keystone-operator-controller-manager-5b84cc7657-k5rmd\" (UID: \"64565126-cba9-4da4-8e0e-7a75f2acc4f1\") " pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-k5rmd" Oct 07 13:59:45 crc kubenswrapper[4959]: E1007 13:59:45.442978 4959 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 07 13:59:45 crc kubenswrapper[4959]: E1007 13:59:45.443015 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/77cf7bcc-63b9-4ee3-84bb-4d4b784750c6-cert podName:77cf7bcc-63b9-4ee3-84bb-4d4b784750c6 nodeName:}" failed. No retries permitted until 2025-10-07 13:59:45.943001308 +0000 UTC m=+868.026405633 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/77cf7bcc-63b9-4ee3-84bb-4d4b784750c6-cert") pod "infra-operator-controller-manager-84788b6bc5-rdj2q" (UID: "77cf7bcc-63b9-4ee3-84bb-4d4b784750c6") : secret "infra-operator-webhook-server-cert" not found Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.449367 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.449659 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-48x54" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.463242 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.470041 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-pnbwt" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.478701 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbwsj\" (UniqueName: \"kubernetes.io/projected/c21b1db6-de7b-48eb-9796-b7d45a638e8c-kube-api-access-gbwsj\") pod \"horizon-operator-controller-manager-5b477879bc-p6mgb\" (UID: \"c21b1db6-de7b-48eb-9796-b7d45a638e8c\") " pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-p6mgb" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.487548 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-264kt\" (UniqueName: \"kubernetes.io/projected/ca1a8a8b-1649-49b3-b595-0e12408673c9-kube-api-access-264kt\") pod \"glance-operator-controller-manager-fd648f65-bc7rh\" (UID: \"ca1a8a8b-1649-49b3-b595-0e12408673c9\") " pod="openstack-operators/glance-operator-controller-manager-fd648f65-bc7rh" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.493931 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8phv\" (UniqueName: \"kubernetes.io/projected/77cf7bcc-63b9-4ee3-84bb-4d4b784750c6-kube-api-access-z8phv\") pod \"infra-operator-controller-manager-84788b6bc5-rdj2q\" (UID: \"77cf7bcc-63b9-4ee3-84bb-4d4b784750c6\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-rdj2q" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.502675 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54d485fd9-7x6df"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.502765 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcpjt\" (UniqueName: \"kubernetes.io/projected/98825e89-a8c2-4023-b139-4c6f5d9b7987-kube-api-access-hcpjt\") pod \"heat-operator-controller-manager-7ccfc8cf49-q56bw\" (UID: \"98825e89-a8c2-4023-b139-4c6f5d9b7987\") " pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-q56bw" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.504371 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-7x6df" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.516582 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-p6mgb" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.517248 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-zmjd4" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.533018 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54d485fd9-7x6df"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.545409 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8g29\" (UniqueName: \"kubernetes.io/projected/e1da143c-64e7-4c75-97fb-4e3553fc2906-kube-api-access-n8g29\") pod \"mariadb-operator-controller-manager-d6c9dc5bc-ksjkq\" (UID: \"e1da143c-64e7-4c75-97fb-4e3553fc2906\") " pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-ksjkq" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.545448 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xckcl\" (UniqueName: \"kubernetes.io/projected/97882e07-7ee2-4b7e-86fc-8033bdd78934-kube-api-access-xckcl\") pod \"nova-operator-controller-manager-6c9b57c67-mgrc2\" (UID: \"97882e07-7ee2-4b7e-86fc-8033bdd78934\") " pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-mgrc2" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.545472 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzfst\" (UniqueName: \"kubernetes.io/projected/9bee8f50-89db-4ddb-a033-6f15137aff40-kube-api-access-zzfst\") pod \"octavia-operator-controller-manager-69f59f9d8-tfrz7\" (UID: \"9bee8f50-89db-4ddb-a033-6f15137aff40\") " pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-tfrz7" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.545498 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxtw8\" (UniqueName: \"kubernetes.io/projected/26ef3538-9145-485b-aa33-f888f7a9d46e-kube-api-access-sxtw8\") pod \"neutron-operator-controller-manager-69b956fbf6-k4pkf\" (UID: \"26ef3538-9145-485b-aa33-f888f7a9d46e\") " pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-k4pkf" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.545517 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dzkl\" (UniqueName: \"kubernetes.io/projected/801942fb-8de2-449b-9a14-4616bdec95aa-kube-api-access-5dzkl\") pod \"manila-operator-controller-manager-7cb48dbc-hhbqh\" (UID: \"801942fb-8de2-449b-9a14-4616bdec95aa\") " pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-hhbqh" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.545553 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdh85\" (UniqueName: \"kubernetes.io/projected/64565126-cba9-4da4-8e0e-7a75f2acc4f1-kube-api-access-vdh85\") pod \"keystone-operator-controller-manager-5b84cc7657-k5rmd\" (UID: \"64565126-cba9-4da4-8e0e-7a75f2acc4f1\") " pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-k5rmd" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.545584 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8k9cn\" (UniqueName: \"kubernetes.io/projected/6450acc6-a3fb-429e-903e-8f99ab864a71-kube-api-access-8k9cn\") pod \"openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9\" (UID: \"6450acc6-a3fb-429e-903e-8f99ab864a71\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.545600 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6450acc6-a3fb-429e-903e-8f99ab864a71-cert\") pod \"openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9\" (UID: \"6450acc6-a3fb-429e-903e-8f99ab864a71\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.545635 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldw5w\" (UniqueName: \"kubernetes.io/projected/c47e0f74-de21-41f6-9142-85d47fcafc60-kube-api-access-ldw5w\") pod \"ironic-operator-controller-manager-5467f8988c-l4hdq\" (UID: \"c47e0f74-de21-41f6-9142-85d47fcafc60\") " pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-l4hdq" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.564343 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-66f6d6849b-zdxb9"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.573038 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dzkl\" (UniqueName: \"kubernetes.io/projected/801942fb-8de2-449b-9a14-4616bdec95aa-kube-api-access-5dzkl\") pod \"manila-operator-controller-manager-7cb48dbc-hhbqh\" (UID: \"801942fb-8de2-449b-9a14-4616bdec95aa\") " pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-hhbqh" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.573485 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldw5w\" (UniqueName: \"kubernetes.io/projected/c47e0f74-de21-41f6-9142-85d47fcafc60-kube-api-access-ldw5w\") pod \"ironic-operator-controller-manager-5467f8988c-l4hdq\" (UID: \"c47e0f74-de21-41f6-9142-85d47fcafc60\") " pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-l4hdq" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.575412 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8g29\" (UniqueName: \"kubernetes.io/projected/e1da143c-64e7-4c75-97fb-4e3553fc2906-kube-api-access-n8g29\") pod \"mariadb-operator-controller-manager-d6c9dc5bc-ksjkq\" (UID: \"e1da143c-64e7-4c75-97fb-4e3553fc2906\") " pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-ksjkq" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.583610 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-76d5577b-4tp8k"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.584351 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdh85\" (UniqueName: \"kubernetes.io/projected/64565126-cba9-4da4-8e0e-7a75f2acc4f1-kube-api-access-vdh85\") pod \"keystone-operator-controller-manager-5b84cc7657-k5rmd\" (UID: \"64565126-cba9-4da4-8e0e-7a75f2acc4f1\") " pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-k5rmd" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.584608 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-76d5577b-4tp8k" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.585016 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-zdxb9" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.589052 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-5j6rs" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.589274 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-98z44" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.609520 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-76d5577b-4tp8k"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.632290 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-66f6d6849b-zdxb9"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.637355 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-f589c7597-lcjnz"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.638495 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-lcjnz" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.643148 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-l4hdq" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.646285 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-f589c7597-lcjnz"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.657500 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxtw8\" (UniqueName: \"kubernetes.io/projected/26ef3538-9145-485b-aa33-f888f7a9d46e-kube-api-access-sxtw8\") pod \"neutron-operator-controller-manager-69b956fbf6-k4pkf\" (UID: \"26ef3538-9145-485b-aa33-f888f7a9d46e\") " pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-k4pkf" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.657586 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8k9cn\" (UniqueName: \"kubernetes.io/projected/6450acc6-a3fb-429e-903e-8f99ab864a71-kube-api-access-8k9cn\") pod \"openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9\" (UID: \"6450acc6-a3fb-429e-903e-8f99ab864a71\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.657645 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6450acc6-a3fb-429e-903e-8f99ab864a71-cert\") pod \"openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9\" (UID: \"6450acc6-a3fb-429e-903e-8f99ab864a71\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.657752 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9zcp\" (UniqueName: \"kubernetes.io/projected/b6e75c33-53ea-4ecb-bb66-1dfb86201e9d-kube-api-access-w9zcp\") pod \"ovn-operator-controller-manager-54d485fd9-7x6df\" (UID: \"b6e75c33-53ea-4ecb-bb66-1dfb86201e9d\") " pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-7x6df" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.657792 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xckcl\" (UniqueName: \"kubernetes.io/projected/97882e07-7ee2-4b7e-86fc-8033bdd78934-kube-api-access-xckcl\") pod \"nova-operator-controller-manager-6c9b57c67-mgrc2\" (UID: \"97882e07-7ee2-4b7e-86fc-8033bdd78934\") " pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-mgrc2" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.658479 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzfst\" (UniqueName: \"kubernetes.io/projected/9bee8f50-89db-4ddb-a033-6f15137aff40-kube-api-access-zzfst\") pod \"octavia-operator-controller-manager-69f59f9d8-tfrz7\" (UID: \"9bee8f50-89db-4ddb-a033-6f15137aff40\") " pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-tfrz7" Oct 07 13:59:45 crc kubenswrapper[4959]: E1007 13:59:45.659027 4959 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 07 13:59:45 crc kubenswrapper[4959]: E1007 13:59:45.659143 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6450acc6-a3fb-429e-903e-8f99ab864a71-cert podName:6450acc6-a3fb-429e-903e-8f99ab864a71 nodeName:}" failed. No retries permitted until 2025-10-07 13:59:46.159115728 +0000 UTC m=+868.242520053 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6450acc6-a3fb-429e-903e-8f99ab864a71-cert") pod "openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9" (UID: "6450acc6-a3fb-429e-903e-8f99ab864a71") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.678198 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-k5rmd" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.689564 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-hhbqh" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.696983 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-4hkp4" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.706595 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-ksjkq" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.731345 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzfst\" (UniqueName: \"kubernetes.io/projected/9bee8f50-89db-4ddb-a033-6f15137aff40-kube-api-access-zzfst\") pod \"octavia-operator-controller-manager-69f59f9d8-tfrz7\" (UID: \"9bee8f50-89db-4ddb-a033-6f15137aff40\") " pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-tfrz7" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.731909 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8k9cn\" (UniqueName: \"kubernetes.io/projected/6450acc6-a3fb-429e-903e-8f99ab864a71-kube-api-access-8k9cn\") pod \"openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9\" (UID: \"6450acc6-a3fb-429e-903e-8f99ab864a71\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.737349 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxtw8\" (UniqueName: \"kubernetes.io/projected/26ef3538-9145-485b-aa33-f888f7a9d46e-kube-api-access-sxtw8\") pod \"neutron-operator-controller-manager-69b956fbf6-k4pkf\" (UID: \"26ef3538-9145-485b-aa33-f888f7a9d46e\") " pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-k4pkf" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.760051 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-6bb6dcddc-nms79"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.761806 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-nms79" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.769764 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52zws\" (UniqueName: \"kubernetes.io/projected/c28ad232-e3aa-425b-a04d-e1f31cd22770-kube-api-access-52zws\") pod \"swift-operator-controller-manager-76d5577b-4tp8k\" (UID: \"c28ad232-e3aa-425b-a04d-e1f31cd22770\") " pod="openstack-operators/swift-operator-controller-manager-76d5577b-4tp8k" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.769823 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9zcp\" (UniqueName: \"kubernetes.io/projected/b6e75c33-53ea-4ecb-bb66-1dfb86201e9d-kube-api-access-w9zcp\") pod \"ovn-operator-controller-manager-54d485fd9-7x6df\" (UID: \"b6e75c33-53ea-4ecb-bb66-1dfb86201e9d\") " pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-7x6df" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.769881 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l6g6\" (UniqueName: \"kubernetes.io/projected/e1e7bf8b-8f59-486f-920e-adfa13c89d2a-kube-api-access-8l6g6\") pod \"placement-operator-controller-manager-66f6d6849b-zdxb9\" (UID: \"e1e7bf8b-8f59-486f-920e-adfa13c89d2a\") " pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-zdxb9" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.769995 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sf59\" (UniqueName: \"kubernetes.io/projected/793bfbb8-e13c-4c1c-a490-d422b79b88c5-kube-api-access-7sf59\") pod \"telemetry-operator-controller-manager-f589c7597-lcjnz\" (UID: \"793bfbb8-e13c-4c1c-a490-d422b79b88c5\") " pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-lcjnz" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.770891 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-fd648f65-bc7rh" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.771352 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-66kkl" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.778520 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-6bb6dcddc-nms79"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.787154 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-q56bw" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.787463 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xckcl\" (UniqueName: \"kubernetes.io/projected/97882e07-7ee2-4b7e-86fc-8033bdd78934-kube-api-access-xckcl\") pod \"nova-operator-controller-manager-6c9b57c67-mgrc2\" (UID: \"97882e07-7ee2-4b7e-86fc-8033bdd78934\") " pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-mgrc2" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.801699 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9zcp\" (UniqueName: \"kubernetes.io/projected/b6e75c33-53ea-4ecb-bb66-1dfb86201e9d-kube-api-access-w9zcp\") pod \"ovn-operator-controller-manager-54d485fd9-7x6df\" (UID: \"b6e75c33-53ea-4ecb-bb66-1dfb86201e9d\") " pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-7x6df" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.821852 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-tfrz7" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.823921 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5d98cc5575-74gv9"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.824956 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-74gv9" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.831662 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-qdswv" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.834247 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5d98cc5575-74gv9"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.838496 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-7x6df" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.874722 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkb64\" (UniqueName: \"kubernetes.io/projected/b875d05e-5d50-423f-9c63-66f7f1c2d163-kube-api-access-kkb64\") pod \"test-operator-controller-manager-6bb6dcddc-nms79\" (UID: \"b875d05e-5d50-423f-9c63-66f7f1c2d163\") " pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-nms79" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.874776 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sf59\" (UniqueName: \"kubernetes.io/projected/793bfbb8-e13c-4c1c-a490-d422b79b88c5-kube-api-access-7sf59\") pod \"telemetry-operator-controller-manager-f589c7597-lcjnz\" (UID: \"793bfbb8-e13c-4c1c-a490-d422b79b88c5\") " pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-lcjnz" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.874818 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52zws\" (UniqueName: \"kubernetes.io/projected/c28ad232-e3aa-425b-a04d-e1f31cd22770-kube-api-access-52zws\") pod \"swift-operator-controller-manager-76d5577b-4tp8k\" (UID: \"c28ad232-e3aa-425b-a04d-e1f31cd22770\") " pod="openstack-operators/swift-operator-controller-manager-76d5577b-4tp8k" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.874865 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8l6g6\" (UniqueName: \"kubernetes.io/projected/e1e7bf8b-8f59-486f-920e-adfa13c89d2a-kube-api-access-8l6g6\") pod \"placement-operator-controller-manager-66f6d6849b-zdxb9\" (UID: \"e1e7bf8b-8f59-486f-920e-adfa13c89d2a\") " pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-zdxb9" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.896995 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-fd79fd9-mt8qw"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.898122 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-fd79fd9-mt8qw" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.914419 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-hrkzm" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.915008 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.920223 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-fd79fd9-mt8qw"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.923234 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8l6g6\" (UniqueName: \"kubernetes.io/projected/e1e7bf8b-8f59-486f-920e-adfa13c89d2a-kube-api-access-8l6g6\") pod \"placement-operator-controller-manager-66f6d6849b-zdxb9\" (UID: \"e1e7bf8b-8f59-486f-920e-adfa13c89d2a\") " pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-zdxb9" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.924793 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sf59\" (UniqueName: \"kubernetes.io/projected/793bfbb8-e13c-4c1c-a490-d422b79b88c5-kube-api-access-7sf59\") pod \"telemetry-operator-controller-manager-f589c7597-lcjnz\" (UID: \"793bfbb8-e13c-4c1c-a490-d422b79b88c5\") " pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-lcjnz" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.926791 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52zws\" (UniqueName: \"kubernetes.io/projected/c28ad232-e3aa-425b-a04d-e1f31cd22770-kube-api-access-52zws\") pod \"swift-operator-controller-manager-76d5577b-4tp8k\" (UID: \"c28ad232-e3aa-425b-a04d-e1f31cd22770\") " pod="openstack-operators/swift-operator-controller-manager-76d5577b-4tp8k" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.936297 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-zdxb9" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.965588 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-99v5c"] Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.966908 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-99v5c" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.975941 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-lcjnz" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.976488 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-p2qwt" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.977579 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkb64\" (UniqueName: \"kubernetes.io/projected/b875d05e-5d50-423f-9c63-66f7f1c2d163-kube-api-access-kkb64\") pod \"test-operator-controller-manager-6bb6dcddc-nms79\" (UID: \"b875d05e-5d50-423f-9c63-66f7f1c2d163\") " pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-nms79" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.977643 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggsr7\" (UniqueName: \"kubernetes.io/projected/de837fcd-bcf9-4582-86e9-806b5442dae9-kube-api-access-ggsr7\") pod \"watcher-operator-controller-manager-5d98cc5575-74gv9\" (UID: \"de837fcd-bcf9-4582-86e9-806b5442dae9\") " pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-74gv9" Oct 07 13:59:45 crc kubenswrapper[4959]: I1007 13:59:45.977676 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/77cf7bcc-63b9-4ee3-84bb-4d4b784750c6-cert\") pod \"infra-operator-controller-manager-84788b6bc5-rdj2q\" (UID: \"77cf7bcc-63b9-4ee3-84bb-4d4b784750c6\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-rdj2q" Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:45.992239 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/77cf7bcc-63b9-4ee3-84bb-4d4b784750c6-cert\") pod \"infra-operator-controller-manager-84788b6bc5-rdj2q\" (UID: \"77cf7bcc-63b9-4ee3-84bb-4d4b784750c6\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-rdj2q" Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:46.026840 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-k4pkf" Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:46.046480 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-mgrc2" Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:46.049450 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-99v5c"] Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:46.069167 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkb64\" (UniqueName: \"kubernetes.io/projected/b875d05e-5d50-423f-9c63-66f7f1c2d163-kube-api-access-kkb64\") pod \"test-operator-controller-manager-6bb6dcddc-nms79\" (UID: \"b875d05e-5d50-423f-9c63-66f7f1c2d163\") " pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-nms79" Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:46.082706 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggsr7\" (UniqueName: \"kubernetes.io/projected/de837fcd-bcf9-4582-86e9-806b5442dae9-kube-api-access-ggsr7\") pod \"watcher-operator-controller-manager-5d98cc5575-74gv9\" (UID: \"de837fcd-bcf9-4582-86e9-806b5442dae9\") " pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-74gv9" Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:46.082753 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b86a7280-99fd-4b0a-bfeb-cab73d5c360a-cert\") pod \"openstack-operator-controller-manager-fd79fd9-mt8qw\" (UID: \"b86a7280-99fd-4b0a-bfeb-cab73d5c360a\") " pod="openstack-operators/openstack-operator-controller-manager-fd79fd9-mt8qw" Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:46.082794 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzf6z\" (UniqueName: \"kubernetes.io/projected/54ba157f-2ce3-4fed-aab7-b9e0c1056b78-kube-api-access-nzf6z\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-99v5c\" (UID: \"54ba157f-2ce3-4fed-aab7-b9e0c1056b78\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-99v5c" Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:46.082873 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhvch\" (UniqueName: \"kubernetes.io/projected/b86a7280-99fd-4b0a-bfeb-cab73d5c360a-kube-api-access-qhvch\") pod \"openstack-operator-controller-manager-fd79fd9-mt8qw\" (UID: \"b86a7280-99fd-4b0a-bfeb-cab73d5c360a\") " pod="openstack-operators/openstack-operator-controller-manager-fd79fd9-mt8qw" Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:46.097979 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-nms79" Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:46.123985 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggsr7\" (UniqueName: \"kubernetes.io/projected/de837fcd-bcf9-4582-86e9-806b5442dae9-kube-api-access-ggsr7\") pod \"watcher-operator-controller-manager-5d98cc5575-74gv9\" (UID: \"de837fcd-bcf9-4582-86e9-806b5442dae9\") " pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-74gv9" Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:46.161011 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-58d86cd59d-pnbwt"] Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:46.162774 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-rdj2q" Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:46.180922 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-74gv9" Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:46.183790 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhvch\" (UniqueName: \"kubernetes.io/projected/b86a7280-99fd-4b0a-bfeb-cab73d5c360a-kube-api-access-qhvch\") pod \"openstack-operator-controller-manager-fd79fd9-mt8qw\" (UID: \"b86a7280-99fd-4b0a-bfeb-cab73d5c360a\") " pod="openstack-operators/openstack-operator-controller-manager-fd79fd9-mt8qw" Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:46.183842 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b86a7280-99fd-4b0a-bfeb-cab73d5c360a-cert\") pod \"openstack-operator-controller-manager-fd79fd9-mt8qw\" (UID: \"b86a7280-99fd-4b0a-bfeb-cab73d5c360a\") " pod="openstack-operators/openstack-operator-controller-manager-fd79fd9-mt8qw" Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:46.183882 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzf6z\" (UniqueName: \"kubernetes.io/projected/54ba157f-2ce3-4fed-aab7-b9e0c1056b78-kube-api-access-nzf6z\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-99v5c\" (UID: \"54ba157f-2ce3-4fed-aab7-b9e0c1056b78\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-99v5c" Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:46.183933 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6450acc6-a3fb-429e-903e-8f99ab864a71-cert\") pod \"openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9\" (UID: \"6450acc6-a3fb-429e-903e-8f99ab864a71\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9" Oct 07 13:59:46 crc kubenswrapper[4959]: E1007 13:59:46.184052 4959 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 07 13:59:46 crc kubenswrapper[4959]: E1007 13:59:46.184112 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6450acc6-a3fb-429e-903e-8f99ab864a71-cert podName:6450acc6-a3fb-429e-903e-8f99ab864a71 nodeName:}" failed. No retries permitted until 2025-10-07 13:59:47.184083683 +0000 UTC m=+869.267488008 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6450acc6-a3fb-429e-903e-8f99ab864a71-cert") pod "openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9" (UID: "6450acc6-a3fb-429e-903e-8f99ab864a71") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 07 13:59:46 crc kubenswrapper[4959]: E1007 13:59:46.184297 4959 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 07 13:59:46 crc kubenswrapper[4959]: E1007 13:59:46.184352 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b86a7280-99fd-4b0a-bfeb-cab73d5c360a-cert podName:b86a7280-99fd-4b0a-bfeb-cab73d5c360a nodeName:}" failed. No retries permitted until 2025-10-07 13:59:46.684338769 +0000 UTC m=+868.767743094 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b86a7280-99fd-4b0a-bfeb-cab73d5c360a-cert") pod "openstack-operator-controller-manager-fd79fd9-mt8qw" (UID: "b86a7280-99fd-4b0a-bfeb-cab73d5c360a") : secret "webhook-server-cert" not found Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:46.184773 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f56ff694-9k7x7"] Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:46.209986 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhvch\" (UniqueName: \"kubernetes.io/projected/b86a7280-99fd-4b0a-bfeb-cab73d5c360a-kube-api-access-qhvch\") pod \"openstack-operator-controller-manager-fd79fd9-mt8qw\" (UID: \"b86a7280-99fd-4b0a-bfeb-cab73d5c360a\") " pod="openstack-operators/openstack-operator-controller-manager-fd79fd9-mt8qw" Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:46.210088 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-76d5577b-4tp8k" Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:46.212242 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzf6z\" (UniqueName: \"kubernetes.io/projected/54ba157f-2ce3-4fed-aab7-b9e0c1056b78-kube-api-access-nzf6z\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-99v5c\" (UID: \"54ba157f-2ce3-4fed-aab7-b9e0c1056b78\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-99v5c" Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:46.454594 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-99v5c" Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:46.490283 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-pnbwt" event={"ID":"e46e007a-5053-443a-9619-247dd3d79a8e","Type":"ContainerStarted","Data":"989c1280830a87ffde1dadfd536fc2691dd845a5f4415efacf3c6be8d62cb7ae"} Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:46.491394 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-9k7x7" event={"ID":"1e3c7b8e-8a44-4539-9e01-4fe9ce17266e","Type":"ContainerStarted","Data":"c59f3c8e5a457c4593eab0598ec65a9d348085e45b6d79a5d1854b28636c8a97"} Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:46.692654 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b86a7280-99fd-4b0a-bfeb-cab73d5c360a-cert\") pod \"openstack-operator-controller-manager-fd79fd9-mt8qw\" (UID: \"b86a7280-99fd-4b0a-bfeb-cab73d5c360a\") " pod="openstack-operators/openstack-operator-controller-manager-fd79fd9-mt8qw" Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:46.697560 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b86a7280-99fd-4b0a-bfeb-cab73d5c360a-cert\") pod \"openstack-operator-controller-manager-fd79fd9-mt8qw\" (UID: \"b86a7280-99fd-4b0a-bfeb-cab73d5c360a\") " pod="openstack-operators/openstack-operator-controller-manager-fd79fd9-mt8qw" Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:46.788739 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b477879bc-p6mgb"] Oct 07 13:59:46 crc kubenswrapper[4959]: W1007 13:59:46.798030 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc21b1db6_de7b_48eb_9796_b7d45a638e8c.slice/crio-f38316e8be5164d635dace972c9b0a812ed11d9073cfd8a835ab48715c71307e WatchSource:0}: Error finding container f38316e8be5164d635dace972c9b0a812ed11d9073cfd8a835ab48715c71307e: Status 404 returned error can't find the container with id f38316e8be5164d635dace972c9b0a812ed11d9073cfd8a835ab48715c71307e Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:46.846973 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-fd79fd9-mt8qw" Oct 07 13:59:46 crc kubenswrapper[4959]: I1007 13:59:46.960262 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-84bd8f6848-85vgl"] Oct 07 13:59:46 crc kubenswrapper[4959]: W1007 13:59:46.990917 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b84227a_34a9_4969_97be_4750eb555f22.slice/crio-68366677dce165e1783688cf799a385f9e16ca3543369600514aa97f26e82dae WatchSource:0}: Error finding container 68366677dce165e1783688cf799a385f9e16ca3543369600514aa97f26e82dae: Status 404 returned error can't find the container with id 68366677dce165e1783688cf799a385f9e16ca3543369600514aa97f26e82dae Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.204723 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6450acc6-a3fb-429e-903e-8f99ab864a71-cert\") pod \"openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9\" (UID: \"6450acc6-a3fb-429e-903e-8f99ab864a71\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9" Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.209891 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6450acc6-a3fb-429e-903e-8f99ab864a71-cert\") pod \"openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9\" (UID: \"6450acc6-a3fb-429e-903e-8f99ab864a71\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9" Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.240155 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-fd648f65-bc7rh"] Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.245346 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7cb48dbc-hhbqh"] Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.279908 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5b84cc7657-k5rmd"] Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.285453 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-ksjkq"] Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.293288 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54d485fd9-7x6df"] Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.310016 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-76d5577b-4tp8k"] Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.322742 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9" Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.323356 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5467f8988c-l4hdq"] Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.333403 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-f589c7597-lcjnz"] Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.346023 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-6c9b57c67-mgrc2"] Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.358387 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-66f6d6849b-zdxb9"] Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.515761 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-76d5577b-4tp8k" event={"ID":"c28ad232-e3aa-425b-a04d-e1f31cd22770","Type":"ContainerStarted","Data":"1bec1ba04369c2a705b3c50379cbdc563915a497106bde1fbade4ad4c1e3a383"} Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.517259 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-fd648f65-bc7rh" event={"ID":"ca1a8a8b-1649-49b3-b595-0e12408673c9","Type":"ContainerStarted","Data":"6a575f8864b74b993fe6da565b656728f4021abee1129f8f3b8b798b4b67da36"} Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.518399 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-mgrc2" event={"ID":"97882e07-7ee2-4b7e-86fc-8033bdd78934","Type":"ContainerStarted","Data":"da2712cbff8eed9695b3385d1627cb95d1b8cba14beea0a5fc681124d5fed6ea"} Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.519228 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-ksjkq" event={"ID":"e1da143c-64e7-4c75-97fb-4e3553fc2906","Type":"ContainerStarted","Data":"8782e46bd79d591d23e1122da2f98dfb723feaa8cab7a835f9fc74c56a03b87c"} Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.519850 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-hhbqh" event={"ID":"801942fb-8de2-449b-9a14-4616bdec95aa","Type":"ContainerStarted","Data":"d203186cd87df437e329d80ad152aaf26e1ad5c8f53e203179dc91ef5f5a78e8"} Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.521028 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-l4hdq" event={"ID":"c47e0f74-de21-41f6-9142-85d47fcafc60","Type":"ContainerStarted","Data":"0377560c00f659381e558043c603e5b6c8a5c7b887b01ce75a42d03ea8620041"} Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.521866 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-p6mgb" event={"ID":"c21b1db6-de7b-48eb-9796-b7d45a638e8c","Type":"ContainerStarted","Data":"f38316e8be5164d635dace972c9b0a812ed11d9073cfd8a835ab48715c71307e"} Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.522527 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-k5rmd" event={"ID":"64565126-cba9-4da4-8e0e-7a75f2acc4f1","Type":"ContainerStarted","Data":"bc4d40ed6c4ef59e5e517e26289317f25387dcc517d8e75e3f4d89764cc3caf8"} Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.523119 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-zdxb9" event={"ID":"e1e7bf8b-8f59-486f-920e-adfa13c89d2a","Type":"ContainerStarted","Data":"3ee00830cd5f4530472f80521b00f5693337f0e3db2aea031e21a0df4471a76c"} Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.525455 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-85vgl" event={"ID":"2b84227a-34a9-4969-97be-4750eb555f22","Type":"ContainerStarted","Data":"68366677dce165e1783688cf799a385f9e16ca3543369600514aa97f26e82dae"} Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.526641 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-7x6df" event={"ID":"b6e75c33-53ea-4ecb-bb66-1dfb86201e9d","Type":"ContainerStarted","Data":"f6a6453424abb722d29949db8a4a5ea414e161a5abc8855e7c7a5f398e28daef"} Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.529946 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-lcjnz" event={"ID":"793bfbb8-e13c-4c1c-a490-d422b79b88c5","Type":"ContainerStarted","Data":"702e42e1b685b6f4e120555ca8cb9748508146327d0ff0f9dcb456bf5d35fbb3"} Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.615230 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-7ccfc8cf49-q56bw"] Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.634121 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-84788b6bc5-rdj2q"] Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.650993 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f59f9d8-tfrz7"] Oct 07 13:59:47 crc kubenswrapper[4959]: E1007 13:59:47.655481 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:b6cef68bfaacdf992a9fa1a6b03a848a48c18cbb6ed12d95561b4b37d858b99f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z8phv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-84788b6bc5-rdj2q_openstack-operators(77cf7bcc-63b9-4ee3-84bb-4d4b784750c6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.656647 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-69b956fbf6-k4pkf"] Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.663484 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-6bb6dcddc-nms79"] Oct 07 13:59:47 crc kubenswrapper[4959]: E1007 13:59:47.676151 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kkb64,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-6bb6dcddc-nms79_openstack-operators(b875d05e-5d50-423f-9c63-66f7f1c2d163): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 07 13:59:47 crc kubenswrapper[4959]: E1007 13:59:47.676415 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:da5c3078d80878d66c616e6f8a0bb909f95d971cde2c612f96fded064113e182,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zzfst,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-69f59f9d8-tfrz7_openstack-operators(9bee8f50-89db-4ddb-a033-6f15137aff40): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.687554 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5d98cc5575-74gv9"] Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.691581 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-99v5c"] Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.694817 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-fd79fd9-mt8qw"] Oct 07 13:59:47 crc kubenswrapper[4959]: E1007 13:59:47.698569 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ggsr7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-5d98cc5575-74gv9_openstack-operators(de837fcd-bcf9-4582-86e9-806b5442dae9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 07 13:59:47 crc kubenswrapper[4959]: W1007 13:59:47.719135 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54ba157f_2ce3_4fed_aab7_b9e0c1056b78.slice/crio-20412a8a0f3f21300a0917ad61340b67ce50e018bc3bf8facee7f784582d6d40 WatchSource:0}: Error finding container 20412a8a0f3f21300a0917ad61340b67ce50e018bc3bf8facee7f784582d6d40: Status 404 returned error can't find the container with id 20412a8a0f3f21300a0917ad61340b67ce50e018bc3bf8facee7f784582d6d40 Oct 07 13:59:47 crc kubenswrapper[4959]: E1007 13:59:47.723303 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nzf6z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-99v5c_openstack-operators(54ba157f-2ce3-4fed-aab7-b9e0c1056b78): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 07 13:59:47 crc kubenswrapper[4959]: W1007 13:59:47.723499 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb86a7280_99fd_4b0a_bfeb_cab73d5c360a.slice/crio-d44a6642a7ecb19e0bd131c9f8b137bed26dbc5452ee1fb201a479d7cd64602e WatchSource:0}: Error finding container d44a6642a7ecb19e0bd131c9f8b137bed26dbc5452ee1fb201a479d7cd64602e: Status 404 returned error can't find the container with id d44a6642a7ecb19e0bd131c9f8b137bed26dbc5452ee1fb201a479d7cd64602e Oct 07 13:59:47 crc kubenswrapper[4959]: E1007 13:59:47.724900 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-99v5c" podUID="54ba157f-2ce3-4fed-aab7-b9e0c1056b78" Oct 07 13:59:47 crc kubenswrapper[4959]: I1007 13:59:47.815616 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9"] Oct 07 13:59:47 crc kubenswrapper[4959]: W1007 13:59:47.857141 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6450acc6_a3fb_429e_903e_8f99ab864a71.slice/crio-e7b31ba1c2e19e76516d025e6199a4df1e96ecd6466a25129b63e628248c50f1 WatchSource:0}: Error finding container e7b31ba1c2e19e76516d025e6199a4df1e96ecd6466a25129b63e628248c50f1: Status 404 returned error can't find the container with id e7b31ba1c2e19e76516d025e6199a4df1e96ecd6466a25129b63e628248c50f1 Oct 07 13:59:48 crc kubenswrapper[4959]: E1007 13:59:48.146790 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-nms79" podUID="b875d05e-5d50-423f-9c63-66f7f1c2d163" Oct 07 13:59:48 crc kubenswrapper[4959]: E1007 13:59:48.147387 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-rdj2q" podUID="77cf7bcc-63b9-4ee3-84bb-4d4b784750c6" Oct 07 13:59:48 crc kubenswrapper[4959]: E1007 13:59:48.176959 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-74gv9" podUID="de837fcd-bcf9-4582-86e9-806b5442dae9" Oct 07 13:59:48 crc kubenswrapper[4959]: E1007 13:59:48.288541 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-tfrz7" podUID="9bee8f50-89db-4ddb-a033-6f15137aff40" Oct 07 13:59:48 crc kubenswrapper[4959]: I1007 13:59:48.559435 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-tfrz7" event={"ID":"9bee8f50-89db-4ddb-a033-6f15137aff40","Type":"ContainerStarted","Data":"9465ac2f26a4a55103fab59cba9d0d710daf5b3fdeb625b3e13654edf50a9f47"} Oct 07 13:59:48 crc kubenswrapper[4959]: I1007 13:59:48.559495 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-tfrz7" event={"ID":"9bee8f50-89db-4ddb-a033-6f15137aff40","Type":"ContainerStarted","Data":"8864148b5716ef32a5700cf09d0c279fed7f5ff3ce77cb4da9f0dd80983357db"} Oct 07 13:59:48 crc kubenswrapper[4959]: E1007 13:59:48.566851 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:da5c3078d80878d66c616e6f8a0bb909f95d971cde2c612f96fded064113e182\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-tfrz7" podUID="9bee8f50-89db-4ddb-a033-6f15137aff40" Oct 07 13:59:48 crc kubenswrapper[4959]: I1007 13:59:48.578705 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-q56bw" event={"ID":"98825e89-a8c2-4023-b139-4c6f5d9b7987","Type":"ContainerStarted","Data":"300d968ca4d12ea2a6ee494da27f4f363f5aee2b2d08d37879b8f4d3e26a64d2"} Oct 07 13:59:48 crc kubenswrapper[4959]: I1007 13:59:48.604548 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-nms79" event={"ID":"b875d05e-5d50-423f-9c63-66f7f1c2d163","Type":"ContainerStarted","Data":"c2b0dd4aef5fb9fe89301676a22a046db6441cc401629e5d243ffaf8fcae95e9"} Oct 07 13:59:48 crc kubenswrapper[4959]: I1007 13:59:48.604599 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-nms79" event={"ID":"b875d05e-5d50-423f-9c63-66f7f1c2d163","Type":"ContainerStarted","Data":"5b04231c1686028ed8e8b2028e67c5f7abc1ce56caf0c1cb7f3410c045e2ec36"} Oct 07 13:59:48 crc kubenswrapper[4959]: E1007 13:59:48.612074 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb\\\"\"" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-nms79" podUID="b875d05e-5d50-423f-9c63-66f7f1c2d163" Oct 07 13:59:48 crc kubenswrapper[4959]: I1007 13:59:48.646047 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-74gv9" event={"ID":"de837fcd-bcf9-4582-86e9-806b5442dae9","Type":"ContainerStarted","Data":"2ae28ed61a81b1852c71d8b078632d2622f4a48b0fada48e19a0e782d8c1d6e1"} Oct 07 13:59:48 crc kubenswrapper[4959]: I1007 13:59:48.646090 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-74gv9" event={"ID":"de837fcd-bcf9-4582-86e9-806b5442dae9","Type":"ContainerStarted","Data":"7057c7e05ab1d280eb333eedeb59e8783ca6f15f5903390a2e0b217d77ae3d47"} Oct 07 13:59:48 crc kubenswrapper[4959]: E1007 13:59:48.648528 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-74gv9" podUID="de837fcd-bcf9-4582-86e9-806b5442dae9" Oct 07 13:59:48 crc kubenswrapper[4959]: I1007 13:59:48.651087 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-99v5c" event={"ID":"54ba157f-2ce3-4fed-aab7-b9e0c1056b78","Type":"ContainerStarted","Data":"20412a8a0f3f21300a0917ad61340b67ce50e018bc3bf8facee7f784582d6d40"} Oct 07 13:59:48 crc kubenswrapper[4959]: E1007 13:59:48.660307 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-99v5c" podUID="54ba157f-2ce3-4fed-aab7-b9e0c1056b78" Oct 07 13:59:48 crc kubenswrapper[4959]: I1007 13:59:48.689564 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-k4pkf" event={"ID":"26ef3538-9145-485b-aa33-f888f7a9d46e","Type":"ContainerStarted","Data":"2a2ef8f5594916a5c9e4ba0dae0f64845a984aa6e519852020f05bdfb6be55aa"} Oct 07 13:59:48 crc kubenswrapper[4959]: I1007 13:59:48.695361 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9" event={"ID":"6450acc6-a3fb-429e-903e-8f99ab864a71","Type":"ContainerStarted","Data":"e7b31ba1c2e19e76516d025e6199a4df1e96ecd6466a25129b63e628248c50f1"} Oct 07 13:59:48 crc kubenswrapper[4959]: I1007 13:59:48.715522 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-rdj2q" event={"ID":"77cf7bcc-63b9-4ee3-84bb-4d4b784750c6","Type":"ContainerStarted","Data":"8d118fe117ef2db1aafa5aabdc7c8ab977b2a9297535830597b73eb60215f006"} Oct 07 13:59:48 crc kubenswrapper[4959]: I1007 13:59:48.715577 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-rdj2q" event={"ID":"77cf7bcc-63b9-4ee3-84bb-4d4b784750c6","Type":"ContainerStarted","Data":"c6374b67b3ec204a74b15bd93a77a62e1b9dc287332b0f61597074691d78f1dd"} Oct 07 13:59:48 crc kubenswrapper[4959]: E1007 13:59:48.721373 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:b6cef68bfaacdf992a9fa1a6b03a848a48c18cbb6ed12d95561b4b37d858b99f\\\"\"" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-rdj2q" podUID="77cf7bcc-63b9-4ee3-84bb-4d4b784750c6" Oct 07 13:59:48 crc kubenswrapper[4959]: I1007 13:59:48.754257 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-fd79fd9-mt8qw" event={"ID":"b86a7280-99fd-4b0a-bfeb-cab73d5c360a","Type":"ContainerStarted","Data":"23a5ae1df2b3fc01f8c6e134324d68ccabe9a132b7374255f1be73eacbd7e64c"} Oct 07 13:59:48 crc kubenswrapper[4959]: I1007 13:59:48.754302 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-fd79fd9-mt8qw" event={"ID":"b86a7280-99fd-4b0a-bfeb-cab73d5c360a","Type":"ContainerStarted","Data":"d44a6642a7ecb19e0bd131c9f8b137bed26dbc5452ee1fb201a479d7cd64602e"} Oct 07 13:59:49 crc kubenswrapper[4959]: I1007 13:59:49.764830 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-fd79fd9-mt8qw" event={"ID":"b86a7280-99fd-4b0a-bfeb-cab73d5c360a","Type":"ContainerStarted","Data":"77564760415e77808c368fcb06726fbf14a92a7af5ea8ad2f768e8aed6aea34f"} Oct 07 13:59:49 crc kubenswrapper[4959]: E1007 13:59:49.766442 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:b6cef68bfaacdf992a9fa1a6b03a848a48c18cbb6ed12d95561b4b37d858b99f\\\"\"" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-rdj2q" podUID="77cf7bcc-63b9-4ee3-84bb-4d4b784750c6" Oct 07 13:59:49 crc kubenswrapper[4959]: E1007 13:59:49.766899 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-74gv9" podUID="de837fcd-bcf9-4582-86e9-806b5442dae9" Oct 07 13:59:49 crc kubenswrapper[4959]: E1007 13:59:49.766921 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:da5c3078d80878d66c616e6f8a0bb909f95d971cde2c612f96fded064113e182\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-tfrz7" podUID="9bee8f50-89db-4ddb-a033-6f15137aff40" Oct 07 13:59:49 crc kubenswrapper[4959]: E1007 13:59:49.767116 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-99v5c" podUID="54ba157f-2ce3-4fed-aab7-b9e0c1056b78" Oct 07 13:59:49 crc kubenswrapper[4959]: E1007 13:59:49.768061 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb\\\"\"" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-nms79" podUID="b875d05e-5d50-423f-9c63-66f7f1c2d163" Oct 07 13:59:49 crc kubenswrapper[4959]: I1007 13:59:49.881333 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-fd79fd9-mt8qw" podStartSLOduration=4.881312423 podStartE2EDuration="4.881312423s" podCreationTimestamp="2025-10-07 13:59:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 13:59:49.875668273 +0000 UTC m=+871.959072598" watchObservedRunningTime="2025-10-07 13:59:49.881312423 +0000 UTC m=+871.964716738" Oct 07 13:59:50 crc kubenswrapper[4959]: I1007 13:59:50.771083 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-fd79fd9-mt8qw" Oct 07 13:59:56 crc kubenswrapper[4959]: I1007 13:59:56.852683 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-fd79fd9-mt8qw" Oct 07 14:00:00 crc kubenswrapper[4959]: I1007 14:00:00.131037 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330760-wn777"] Oct 07 14:00:00 crc kubenswrapper[4959]: I1007 14:00:00.132514 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330760-wn777" Oct 07 14:00:00 crc kubenswrapper[4959]: I1007 14:00:00.134249 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 07 14:00:00 crc kubenswrapper[4959]: I1007 14:00:00.134525 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 07 14:00:00 crc kubenswrapper[4959]: I1007 14:00:00.149901 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330760-wn777"] Oct 07 14:00:00 crc kubenswrapper[4959]: I1007 14:00:00.205747 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/444c0979-8132-4868-bd91-5df3278b7660-config-volume\") pod \"collect-profiles-29330760-wn777\" (UID: \"444c0979-8132-4868-bd91-5df3278b7660\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330760-wn777" Oct 07 14:00:00 crc kubenswrapper[4959]: I1007 14:00:00.205846 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fthz5\" (UniqueName: \"kubernetes.io/projected/444c0979-8132-4868-bd91-5df3278b7660-kube-api-access-fthz5\") pod \"collect-profiles-29330760-wn777\" (UID: \"444c0979-8132-4868-bd91-5df3278b7660\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330760-wn777" Oct 07 14:00:00 crc kubenswrapper[4959]: I1007 14:00:00.206093 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/444c0979-8132-4868-bd91-5df3278b7660-secret-volume\") pod \"collect-profiles-29330760-wn777\" (UID: \"444c0979-8132-4868-bd91-5df3278b7660\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330760-wn777" Oct 07 14:00:00 crc kubenswrapper[4959]: I1007 14:00:00.307256 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/444c0979-8132-4868-bd91-5df3278b7660-config-volume\") pod \"collect-profiles-29330760-wn777\" (UID: \"444c0979-8132-4868-bd91-5df3278b7660\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330760-wn777" Oct 07 14:00:00 crc kubenswrapper[4959]: I1007 14:00:00.307577 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fthz5\" (UniqueName: \"kubernetes.io/projected/444c0979-8132-4868-bd91-5df3278b7660-kube-api-access-fthz5\") pod \"collect-profiles-29330760-wn777\" (UID: \"444c0979-8132-4868-bd91-5df3278b7660\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330760-wn777" Oct 07 14:00:00 crc kubenswrapper[4959]: I1007 14:00:00.307634 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/444c0979-8132-4868-bd91-5df3278b7660-secret-volume\") pod \"collect-profiles-29330760-wn777\" (UID: \"444c0979-8132-4868-bd91-5df3278b7660\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330760-wn777" Oct 07 14:00:00 crc kubenswrapper[4959]: I1007 14:00:00.308272 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/444c0979-8132-4868-bd91-5df3278b7660-config-volume\") pod \"collect-profiles-29330760-wn777\" (UID: \"444c0979-8132-4868-bd91-5df3278b7660\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330760-wn777" Oct 07 14:00:00 crc kubenswrapper[4959]: I1007 14:00:00.313880 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/444c0979-8132-4868-bd91-5df3278b7660-secret-volume\") pod \"collect-profiles-29330760-wn777\" (UID: \"444c0979-8132-4868-bd91-5df3278b7660\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330760-wn777" Oct 07 14:00:00 crc kubenswrapper[4959]: I1007 14:00:00.325589 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fthz5\" (UniqueName: \"kubernetes.io/projected/444c0979-8132-4868-bd91-5df3278b7660-kube-api-access-fthz5\") pod \"collect-profiles-29330760-wn777\" (UID: \"444c0979-8132-4868-bd91-5df3278b7660\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330760-wn777" Oct 07 14:00:00 crc kubenswrapper[4959]: I1007 14:00:00.464885 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330760-wn777" Oct 07 14:00:04 crc kubenswrapper[4959]: E1007 14:00:04.091405 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e" Oct 07 14:00:04 crc kubenswrapper[4959]: E1007 14:00:04.091827 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7sf59,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-f589c7597-lcjnz_openstack-operators(793bfbb8-e13c-4c1c-a490-d422b79b88c5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 14:00:10 crc kubenswrapper[4959]: E1007 14:00:10.542341 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-lcjnz" podUID="793bfbb8-e13c-4c1c-a490-d422b79b88c5" Oct 07 14:00:10 crc kubenswrapper[4959]: I1007 14:00:10.665692 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330760-wn777"] Oct 07 14:00:10 crc kubenswrapper[4959]: I1007 14:00:10.926454 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-76d5577b-4tp8k" event={"ID":"c28ad232-e3aa-425b-a04d-e1f31cd22770","Type":"ContainerStarted","Data":"bd569865b70a8d529811ae12812b25caab4185c40fe5b871366d2bf9e3ec419d"} Oct 07 14:00:10 crc kubenswrapper[4959]: I1007 14:00:10.927460 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-mgrc2" event={"ID":"97882e07-7ee2-4b7e-86fc-8033bdd78934","Type":"ContainerStarted","Data":"988ae1e95898739dc4d5e2bb01c6824b2108944f38635096dc7b280b9b54be21"} Oct 07 14:00:10 crc kubenswrapper[4959]: I1007 14:00:10.928477 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-85vgl" event={"ID":"2b84227a-34a9-4969-97be-4750eb555f22","Type":"ContainerStarted","Data":"31520508f8c6dfb216f61acbb5061ea6ec7bbaad88d622730004a45626ba877c"} Oct 07 14:00:10 crc kubenswrapper[4959]: I1007 14:00:10.932431 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-hhbqh" event={"ID":"801942fb-8de2-449b-9a14-4616bdec95aa","Type":"ContainerStarted","Data":"ddd64c8591f96ed49069f1fcf5db88e5956f993713488568024ec45fd3d13fb2"} Oct 07 14:00:10 crc kubenswrapper[4959]: I1007 14:00:10.934309 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-7x6df" event={"ID":"b6e75c33-53ea-4ecb-bb66-1dfb86201e9d","Type":"ContainerStarted","Data":"8698da8188cddb684da138fc37eeb58b17264e5e0dc640d60319ff72384532e4"} Oct 07 14:00:10 crc kubenswrapper[4959]: I1007 14:00:10.936888 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-lcjnz" event={"ID":"793bfbb8-e13c-4c1c-a490-d422b79b88c5","Type":"ContainerStarted","Data":"f678f959966fc4aebf197b540a199a311e80178e21814161c8e4ee15298fa7ca"} Oct 07 14:00:10 crc kubenswrapper[4959]: I1007 14:00:10.941123 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-p6mgb" event={"ID":"c21b1db6-de7b-48eb-9796-b7d45a638e8c","Type":"ContainerStarted","Data":"f163c1606fd4f234f1bd809d39cd6634fd9fd7dd47cf259aa4a8f7b9df4f2c2a"} Oct 07 14:00:10 crc kubenswrapper[4959]: I1007 14:00:10.944402 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330760-wn777" event={"ID":"444c0979-8132-4868-bd91-5df3278b7660","Type":"ContainerStarted","Data":"0c9fff9615e8c94ccbd13e072b777f0327eb2fdd47b46efdd8703bca8013539f"} Oct 07 14:00:10 crc kubenswrapper[4959]: E1007 14:00:10.966467 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-lcjnz" podUID="793bfbb8-e13c-4c1c-a490-d422b79b88c5" Oct 07 14:00:10 crc kubenswrapper[4959]: I1007 14:00:10.966575 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-k4pkf" event={"ID":"26ef3538-9145-485b-aa33-f888f7a9d46e","Type":"ContainerStarted","Data":"1802b236b096f5de4f278c91d8428710d3d58d5db6d27699c7b881ce367cf9fd"} Oct 07 14:00:10 crc kubenswrapper[4959]: I1007 14:00:10.973429 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-pnbwt" event={"ID":"e46e007a-5053-443a-9619-247dd3d79a8e","Type":"ContainerStarted","Data":"e1b6c5c2225327bf44cc63fd7e80d13c345ed7eb5baff219763ff4bbae4bcc16"} Oct 07 14:00:11 crc kubenswrapper[4959]: I1007 14:00:11.981289 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-mgrc2" event={"ID":"97882e07-7ee2-4b7e-86fc-8033bdd78934","Type":"ContainerStarted","Data":"b137b9451799fd66493eeafcb73d9738e346e194ad8f7bbffdf8ff6de6ec63d6"} Oct 07 14:00:11 crc kubenswrapper[4959]: I1007 14:00:11.981854 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-mgrc2" Oct 07 14:00:11 crc kubenswrapper[4959]: I1007 14:00:11.983180 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-85vgl" event={"ID":"2b84227a-34a9-4969-97be-4750eb555f22","Type":"ContainerStarted","Data":"7001765a75e6306a709df043e0107533bc225c693a735d2410659ec3f214112f"} Oct 07 14:00:11 crc kubenswrapper[4959]: I1007 14:00:11.983760 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-85vgl" Oct 07 14:00:11 crc kubenswrapper[4959]: I1007 14:00:11.985752 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-k5rmd" event={"ID":"64565126-cba9-4da4-8e0e-7a75f2acc4f1","Type":"ContainerStarted","Data":"e74ee36d7b948153d06e4be8039f593cee6c34f8a905a15992597cd7b7994c9a"} Oct 07 14:00:11 crc kubenswrapper[4959]: I1007 14:00:11.987186 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-nms79" event={"ID":"b875d05e-5d50-423f-9c63-66f7f1c2d163","Type":"ContainerStarted","Data":"cf263d358fab0aa2059ef810b248612afab31b92c1940599c4975776198a917b"} Oct 07 14:00:11 crc kubenswrapper[4959]: I1007 14:00:11.987543 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-nms79" Oct 07 14:00:11 crc kubenswrapper[4959]: I1007 14:00:11.988477 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-fd648f65-bc7rh" event={"ID":"ca1a8a8b-1649-49b3-b595-0e12408673c9","Type":"ContainerStarted","Data":"f15efff40d7248e3d84679807b552e0da71656c7dac60fdae8d5dc808dd19886"} Oct 07 14:00:11 crc kubenswrapper[4959]: I1007 14:00:11.989546 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-9k7x7" event={"ID":"1e3c7b8e-8a44-4539-9e01-4fe9ce17266e","Type":"ContainerStarted","Data":"71e3f6d352f35082fe0cff585c4b2b8a1f777822d1595aac63f6beaf4dc56348"} Oct 07 14:00:11 crc kubenswrapper[4959]: I1007 14:00:11.990485 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-l4hdq" event={"ID":"c47e0f74-de21-41f6-9142-85d47fcafc60","Type":"ContainerStarted","Data":"747375930b4e8b66cefecf8eb040140681605db10b150ac484ac6f3c541ea5d5"} Oct 07 14:00:11 crc kubenswrapper[4959]: I1007 14:00:11.991906 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9" event={"ID":"6450acc6-a3fb-429e-903e-8f99ab864a71","Type":"ContainerStarted","Data":"052d245dee6767b9892de5d20585f5cf20f49f8ccca4995b656e87f06ed711b0"} Oct 07 14:00:11 crc kubenswrapper[4959]: I1007 14:00:11.997378 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-rdj2q" event={"ID":"77cf7bcc-63b9-4ee3-84bb-4d4b784750c6","Type":"ContainerStarted","Data":"b7d16f8765a97807834c8471262d824903a08e7ef8eee5d959c9e4c171f4518a"} Oct 07 14:00:11 crc kubenswrapper[4959]: I1007 14:00:11.997905 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-rdj2q" Oct 07 14:00:12 crc kubenswrapper[4959]: I1007 14:00:12.002286 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-99v5c" event={"ID":"54ba157f-2ce3-4fed-aab7-b9e0c1056b78","Type":"ContainerStarted","Data":"2aef04500b0d6d2a75bd4ee2b5a91abba1e0d15fee973e9482f539c81255cc0e"} Oct 07 14:00:12 crc kubenswrapper[4959]: I1007 14:00:12.003900 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-76d5577b-4tp8k" event={"ID":"c28ad232-e3aa-425b-a04d-e1f31cd22770","Type":"ContainerStarted","Data":"e60c742c0a137c9e373be82f3cabd77e54fbc5b81989a93722a2c00aeed32966"} Oct 07 14:00:12 crc kubenswrapper[4959]: I1007 14:00:12.004344 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-76d5577b-4tp8k" Oct 07 14:00:12 crc kubenswrapper[4959]: I1007 14:00:12.007975 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-ksjkq" event={"ID":"e1da143c-64e7-4c75-97fb-4e3553fc2906","Type":"ContainerStarted","Data":"04b645a4d2295445be96f018d2bdf8b8c27a5788102042fc040a0767f8993be2"} Oct 07 14:00:12 crc kubenswrapper[4959]: I1007 14:00:12.008465 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-mgrc2" podStartSLOduration=8.598391985 podStartE2EDuration="27.008452678s" podCreationTimestamp="2025-10-07 13:59:45 +0000 UTC" firstStartedPulling="2025-10-07 13:59:47.357432587 +0000 UTC m=+869.440836912" lastFinishedPulling="2025-10-07 14:00:05.76749328 +0000 UTC m=+887.850897605" observedRunningTime="2025-10-07 14:00:12.007715549 +0000 UTC m=+894.091119874" watchObservedRunningTime="2025-10-07 14:00:12.008452678 +0000 UTC m=+894.091857003" Oct 07 14:00:12 crc kubenswrapper[4959]: I1007 14:00:12.009703 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-74gv9" event={"ID":"de837fcd-bcf9-4582-86e9-806b5442dae9","Type":"ContainerStarted","Data":"2d45fd4fb11d90c163a2adce0ff80b7800c01f5a742ecc0e59e354a8c7eafc43"} Oct 07 14:00:12 crc kubenswrapper[4959]: I1007 14:00:12.010076 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-74gv9" Oct 07 14:00:12 crc kubenswrapper[4959]: I1007 14:00:12.011332 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-zdxb9" event={"ID":"e1e7bf8b-8f59-486f-920e-adfa13c89d2a","Type":"ContainerStarted","Data":"796c45f04233bd3faac6c0d7cfe899752678a6e4843d46553ea5e7d8a893620d"} Oct 07 14:00:12 crc kubenswrapper[4959]: I1007 14:00:12.012507 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-q56bw" event={"ID":"98825e89-a8c2-4023-b139-4c6f5d9b7987","Type":"ContainerStarted","Data":"90202b99a515038feb4c4d31aeef513d61fe449fb921a712b64272faef871cf9"} Oct 07 14:00:12 crc kubenswrapper[4959]: E1007 14:00:12.013698 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-lcjnz" podUID="793bfbb8-e13c-4c1c-a490-d422b79b88c5" Oct 07 14:00:12 crc kubenswrapper[4959]: I1007 14:00:12.036825 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-rdj2q" podStartSLOduration=4.324857248 podStartE2EDuration="27.036803829s" podCreationTimestamp="2025-10-07 13:59:45 +0000 UTC" firstStartedPulling="2025-10-07 13:59:47.655345041 +0000 UTC m=+869.738749366" lastFinishedPulling="2025-10-07 14:00:10.367291622 +0000 UTC m=+892.450695947" observedRunningTime="2025-10-07 14:00:12.031720233 +0000 UTC m=+894.115124558" watchObservedRunningTime="2025-10-07 14:00:12.036803829 +0000 UTC m=+894.120208154" Oct 07 14:00:12 crc kubenswrapper[4959]: I1007 14:00:12.047175 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-85vgl" podStartSLOduration=15.234258156 podStartE2EDuration="27.047162326s" podCreationTimestamp="2025-10-07 13:59:45 +0000 UTC" firstStartedPulling="2025-10-07 13:59:46.993272754 +0000 UTC m=+869.076677079" lastFinishedPulling="2025-10-07 13:59:58.806176914 +0000 UTC m=+880.889581249" observedRunningTime="2025-10-07 14:00:12.045384862 +0000 UTC m=+894.128789187" watchObservedRunningTime="2025-10-07 14:00:12.047162326 +0000 UTC m=+894.130566651" Oct 07 14:00:12 crc kubenswrapper[4959]: I1007 14:00:12.066292 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-nms79" podStartSLOduration=4.222385852 podStartE2EDuration="27.066277029s" podCreationTimestamp="2025-10-07 13:59:45 +0000 UTC" firstStartedPulling="2025-10-07 13:59:47.676045564 +0000 UTC m=+869.759449889" lastFinishedPulling="2025-10-07 14:00:10.519936741 +0000 UTC m=+892.603341066" observedRunningTime="2025-10-07 14:00:12.061938281 +0000 UTC m=+894.145342606" watchObservedRunningTime="2025-10-07 14:00:12.066277029 +0000 UTC m=+894.149681354" Oct 07 14:00:12 crc kubenswrapper[4959]: I1007 14:00:12.093665 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-76d5577b-4tp8k" podStartSLOduration=6.455421198 podStartE2EDuration="27.093649136s" podCreationTimestamp="2025-10-07 13:59:45 +0000 UTC" firstStartedPulling="2025-10-07 13:59:47.343567334 +0000 UTC m=+869.426971659" lastFinishedPulling="2025-10-07 14:00:07.981795272 +0000 UTC m=+890.065199597" observedRunningTime="2025-10-07 14:00:12.089967555 +0000 UTC m=+894.173371880" watchObservedRunningTime="2025-10-07 14:00:12.093649136 +0000 UTC m=+894.177053461" Oct 07 14:00:12 crc kubenswrapper[4959]: I1007 14:00:12.115297 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-99v5c" podStartSLOduration=4.330119209 podStartE2EDuration="27.115279392s" podCreationTimestamp="2025-10-07 13:59:45 +0000 UTC" firstStartedPulling="2025-10-07 13:59:47.723194431 +0000 UTC m=+869.806598756" lastFinishedPulling="2025-10-07 14:00:10.508354604 +0000 UTC m=+892.591758939" observedRunningTime="2025-10-07 14:00:12.110278588 +0000 UTC m=+894.193682913" watchObservedRunningTime="2025-10-07 14:00:12.115279392 +0000 UTC m=+894.198683717" Oct 07 14:00:12 crc kubenswrapper[4959]: I1007 14:00:12.138388 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-74gv9" podStartSLOduration=4.469633143 podStartE2EDuration="27.138373104s" podCreationTimestamp="2025-10-07 13:59:45 +0000 UTC" firstStartedPulling="2025-10-07 13:59:47.698461779 +0000 UTC m=+869.781866104" lastFinishedPulling="2025-10-07 14:00:10.36720174 +0000 UTC m=+892.450606065" observedRunningTime="2025-10-07 14:00:12.13701661 +0000 UTC m=+894.220420935" watchObservedRunningTime="2025-10-07 14:00:12.138373104 +0000 UTC m=+894.221777429" Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.019921 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-tfrz7" event={"ID":"9bee8f50-89db-4ddb-a033-6f15137aff40","Type":"ContainerStarted","Data":"46c20897aa85501a1c17302477bae9e21c7a1c51bb520eb6a07c0eec6d9b659d"} Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.020412 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-tfrz7" Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.021690 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-ksjkq" event={"ID":"e1da143c-64e7-4c75-97fb-4e3553fc2906","Type":"ContainerStarted","Data":"62f4c88e8a0a242430b15a2e68c41ece7bd9808da4759dd5c5706b7f60182273"} Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.021790 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-ksjkq" Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.023648 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-fd648f65-bc7rh" event={"ID":"ca1a8a8b-1649-49b3-b595-0e12408673c9","Type":"ContainerStarted","Data":"556dc175e69133864595dcc1616899e5b6f99bf7adc65922299333657e0c7321"} Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.023770 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-fd648f65-bc7rh" Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.025374 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-zdxb9" event={"ID":"e1e7bf8b-8f59-486f-920e-adfa13c89d2a","Type":"ContainerStarted","Data":"971809f7503564f47e8a57c9171a08f9dbebb5152f0f13e3bb3b0baf378296ac"} Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.025524 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-zdxb9" Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.027048 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-k4pkf" event={"ID":"26ef3538-9145-485b-aa33-f888f7a9d46e","Type":"ContainerStarted","Data":"898c19cbd27855a0a0c28b015a0a070d84202955b71a5c5bb8e5aea8372640b7"} Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.027145 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-k4pkf" Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.028737 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-l4hdq" event={"ID":"c47e0f74-de21-41f6-9142-85d47fcafc60","Type":"ContainerStarted","Data":"1d879f49135b0379ca957ad642fcc5ff5bd9b1d5fb933e5b5a19644e78ec1c89"} Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.028844 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-l4hdq" Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.030453 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-q56bw" event={"ID":"98825e89-a8c2-4023-b139-4c6f5d9b7987","Type":"ContainerStarted","Data":"6acef55849d165da19d7772fe32baaf4c3868c1c4df8a0c7b19b99e45c06d54b"} Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.030768 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-q56bw" Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.032149 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-pnbwt" event={"ID":"e46e007a-5053-443a-9619-247dd3d79a8e","Type":"ContainerStarted","Data":"fc2b0232b98c9f515e98849f850b20c5c715c17e2284e4569a89b1ea11f11051"} Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.032296 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-pnbwt" Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.033791 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-hhbqh" event={"ID":"801942fb-8de2-449b-9a14-4616bdec95aa","Type":"ContainerStarted","Data":"2d87eea3ee658d0c49967f540021b99ff7d45b831f65c8d42f9d9001fe89d761"} Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.033978 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-hhbqh" Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.035213 4959 generic.go:334] "Generic (PLEG): container finished" podID="444c0979-8132-4868-bd91-5df3278b7660" containerID="0dab26b0e253a02badce41d64564795611766ab30756fd4692ff9980259c9138" exitCode=0 Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.035282 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330760-wn777" event={"ID":"444c0979-8132-4868-bd91-5df3278b7660","Type":"ContainerDied","Data":"0dab26b0e253a02badce41d64564795611766ab30756fd4692ff9980259c9138"} Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.036932 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-7x6df" event={"ID":"b6e75c33-53ea-4ecb-bb66-1dfb86201e9d","Type":"ContainerStarted","Data":"7f743ec243f48aaac92a788a5ef2d20a3e9e0a77cb42fa338b1b18ae46f016b9"} Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.037000 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-7x6df" Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.039369 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-k5rmd" event={"ID":"64565126-cba9-4da4-8e0e-7a75f2acc4f1","Type":"ContainerStarted","Data":"618bc07f44c76a2efb0e633a334e23f9df9a79a573a24d315efb253133f16112"} Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.039481 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-k5rmd" Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.041058 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-9k7x7" event={"ID":"1e3c7b8e-8a44-4539-9e01-4fe9ce17266e","Type":"ContainerStarted","Data":"e296157054dcae7c192d35c61f4d82b258489cfadf1306d0d2fc6acaa510f7d6"} Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.041209 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-9k7x7" Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.046671 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9" event={"ID":"6450acc6-a3fb-429e-903e-8f99ab864a71","Type":"ContainerStarted","Data":"a404da0ee90272c1fcf6b480e0d0197ba27667ed6afbbe68a7e1a8c2bfc280ea"} Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.046788 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9" Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.048398 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-tfrz7" podStartSLOduration=5.15714064 podStartE2EDuration="28.048387869s" podCreationTimestamp="2025-10-07 13:59:45 +0000 UTC" firstStartedPulling="2025-10-07 13:59:47.676248319 +0000 UTC m=+869.759652644" lastFinishedPulling="2025-10-07 14:00:10.567495548 +0000 UTC m=+892.650899873" observedRunningTime="2025-10-07 14:00:13.045503718 +0000 UTC m=+895.128908063" watchObservedRunningTime="2025-10-07 14:00:13.048387869 +0000 UTC m=+895.131792194" Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.049685 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-p6mgb" event={"ID":"c21b1db6-de7b-48eb-9796-b7d45a638e8c","Type":"ContainerStarted","Data":"ce99e5e291ec2a3bf1f7adb2fb5d9a9c7ee0161ecdc2cd51c7d335a845e675c0"} Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.050270 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-p6mgb" Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.066963 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-ksjkq" podStartSLOduration=7.409137294 podStartE2EDuration="28.066941438s" podCreationTimestamp="2025-10-07 13:59:45 +0000 UTC" firstStartedPulling="2025-10-07 13:59:47.323480146 +0000 UTC m=+869.406884471" lastFinishedPulling="2025-10-07 14:00:07.98128429 +0000 UTC m=+890.064688615" observedRunningTime="2025-10-07 14:00:13.059375151 +0000 UTC m=+895.142779476" watchObservedRunningTime="2025-10-07 14:00:13.066941438 +0000 UTC m=+895.150345763" Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.099528 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-pnbwt" podStartSLOduration=9.798845178 podStartE2EDuration="28.099507134s" podCreationTimestamp="2025-10-07 13:59:45 +0000 UTC" firstStartedPulling="2025-10-07 13:59:46.229993909 +0000 UTC m=+868.313398224" lastFinishedPulling="2025-10-07 14:00:04.530655855 +0000 UTC m=+886.614060180" observedRunningTime="2025-10-07 14:00:13.095553477 +0000 UTC m=+895.178957802" watchObservedRunningTime="2025-10-07 14:00:13.099507134 +0000 UTC m=+895.182911459" Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.102230 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-q56bw" podStartSLOduration=5.49815703 podStartE2EDuration="28.102216121s" podCreationTimestamp="2025-10-07 13:59:45 +0000 UTC" firstStartedPulling="2025-10-07 13:59:47.648802339 +0000 UTC m=+869.732206664" lastFinishedPulling="2025-10-07 14:00:10.25286143 +0000 UTC m=+892.336265755" observedRunningTime="2025-10-07 14:00:13.08113659 +0000 UTC m=+895.164540935" watchObservedRunningTime="2025-10-07 14:00:13.102216121 +0000 UTC m=+895.185620446" Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.111725 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-7x6df" podStartSLOduration=5.466924308 podStartE2EDuration="28.111704716s" podCreationTimestamp="2025-10-07 13:59:45 +0000 UTC" firstStartedPulling="2025-10-07 13:59:47.343536113 +0000 UTC m=+869.426940438" lastFinishedPulling="2025-10-07 14:00:09.988316521 +0000 UTC m=+892.071720846" observedRunningTime="2025-10-07 14:00:13.11062876 +0000 UTC m=+895.194033085" watchObservedRunningTime="2025-10-07 14:00:13.111704716 +0000 UTC m=+895.195109061" Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.140289 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-l4hdq" podStartSLOduration=7.50147295 podStartE2EDuration="28.140272953s" podCreationTimestamp="2025-10-07 13:59:45 +0000 UTC" firstStartedPulling="2025-10-07 13:59:47.343740678 +0000 UTC m=+869.427145003" lastFinishedPulling="2025-10-07 14:00:07.982540681 +0000 UTC m=+890.065945006" observedRunningTime="2025-10-07 14:00:13.139161006 +0000 UTC m=+895.222565331" watchObservedRunningTime="2025-10-07 14:00:13.140272953 +0000 UTC m=+895.223677278" Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.156890 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-fd648f65-bc7rh" podStartSLOduration=8.698664575 podStartE2EDuration="28.156872684s" podCreationTimestamp="2025-10-07 13:59:45 +0000 UTC" firstStartedPulling="2025-10-07 13:59:47.258253462 +0000 UTC m=+869.341657787" lastFinishedPulling="2025-10-07 14:00:06.716461571 +0000 UTC m=+888.799865896" observedRunningTime="2025-10-07 14:00:13.156227688 +0000 UTC m=+895.239632033" watchObservedRunningTime="2025-10-07 14:00:13.156872684 +0000 UTC m=+895.240277009" Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.183718 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-hhbqh" podStartSLOduration=9.638635254 podStartE2EDuration="28.183699999s" podCreationTimestamp="2025-10-07 13:59:45 +0000 UTC" firstStartedPulling="2025-10-07 13:59:47.260070507 +0000 UTC m=+869.343474832" lastFinishedPulling="2025-10-07 14:00:05.805135252 +0000 UTC m=+887.888539577" observedRunningTime="2025-10-07 14:00:13.179221938 +0000 UTC m=+895.262626263" watchObservedRunningTime="2025-10-07 14:00:13.183699999 +0000 UTC m=+895.267104324" Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.200594 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-zdxb9" podStartSLOduration=7.58551284 podStartE2EDuration="28.200575696s" podCreationTimestamp="2025-10-07 13:59:45 +0000 UTC" firstStartedPulling="2025-10-07 13:59:47.366222414 +0000 UTC m=+869.449626739" lastFinishedPulling="2025-10-07 14:00:07.98128527 +0000 UTC m=+890.064689595" observedRunningTime="2025-10-07 14:00:13.197034279 +0000 UTC m=+895.280438604" watchObservedRunningTime="2025-10-07 14:00:13.200575696 +0000 UTC m=+895.283980021" Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.224538 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-9k7x7" podStartSLOduration=7.82460371 podStartE2EDuration="28.224522319s" podCreationTimestamp="2025-10-07 13:59:45 +0000 UTC" firstStartedPulling="2025-10-07 13:59:46.31687099 +0000 UTC m=+868.400275315" lastFinishedPulling="2025-10-07 14:00:06.716789589 +0000 UTC m=+888.800193924" observedRunningTime="2025-10-07 14:00:13.216543682 +0000 UTC m=+895.299948007" watchObservedRunningTime="2025-10-07 14:00:13.224522319 +0000 UTC m=+895.307926644" Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.236317 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-k5rmd" podStartSLOduration=9.755055275 podStartE2EDuration="28.23629687s" podCreationTimestamp="2025-10-07 13:59:45 +0000 UTC" firstStartedPulling="2025-10-07 13:59:47.323826765 +0000 UTC m=+869.407231090" lastFinishedPulling="2025-10-07 14:00:05.80506836 +0000 UTC m=+887.888472685" observedRunningTime="2025-10-07 14:00:13.231790689 +0000 UTC m=+895.315195014" watchObservedRunningTime="2025-10-07 14:00:13.23629687 +0000 UTC m=+895.319701195" Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.250027 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-k4pkf" podStartSLOduration=11.395242894999999 podStartE2EDuration="28.25000912s" podCreationTimestamp="2025-10-07 13:59:45 +0000 UTC" firstStartedPulling="2025-10-07 13:59:47.675813868 +0000 UTC m=+869.759218193" lastFinishedPulling="2025-10-07 14:00:04.530580093 +0000 UTC m=+886.613984418" observedRunningTime="2025-10-07 14:00:13.247793695 +0000 UTC m=+895.331198020" watchObservedRunningTime="2025-10-07 14:00:13.25000912 +0000 UTC m=+895.333413445" Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.262605 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-p6mgb" podStartSLOduration=8.346474248 podStartE2EDuration="28.262591351s" podCreationTimestamp="2025-10-07 13:59:45 +0000 UTC" firstStartedPulling="2025-10-07 13:59:46.800512302 +0000 UTC m=+868.883916627" lastFinishedPulling="2025-10-07 14:00:06.716629385 +0000 UTC m=+888.800033730" observedRunningTime="2025-10-07 14:00:13.260676494 +0000 UTC m=+895.344080819" watchObservedRunningTime="2025-10-07 14:00:13.262591351 +0000 UTC m=+895.345995676" Oct 07 14:00:13 crc kubenswrapper[4959]: I1007 14:00:13.287279 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9" podStartSLOduration=7.292291333 podStartE2EDuration="28.287259972s" podCreationTimestamp="2025-10-07 13:59:45 +0000 UTC" firstStartedPulling="2025-10-07 13:59:47.861653578 +0000 UTC m=+869.945057893" lastFinishedPulling="2025-10-07 14:00:08.856622187 +0000 UTC m=+890.940026532" observedRunningTime="2025-10-07 14:00:13.283114589 +0000 UTC m=+895.366518914" watchObservedRunningTime="2025-10-07 14:00:13.287259972 +0000 UTC m=+895.370664297" Oct 07 14:00:14 crc kubenswrapper[4959]: I1007 14:00:14.433923 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330760-wn777" Oct 07 14:00:14 crc kubenswrapper[4959]: I1007 14:00:14.524809 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fthz5\" (UniqueName: \"kubernetes.io/projected/444c0979-8132-4868-bd91-5df3278b7660-kube-api-access-fthz5\") pod \"444c0979-8132-4868-bd91-5df3278b7660\" (UID: \"444c0979-8132-4868-bd91-5df3278b7660\") " Oct 07 14:00:14 crc kubenswrapper[4959]: I1007 14:00:14.524889 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/444c0979-8132-4868-bd91-5df3278b7660-config-volume\") pod \"444c0979-8132-4868-bd91-5df3278b7660\" (UID: \"444c0979-8132-4868-bd91-5df3278b7660\") " Oct 07 14:00:14 crc kubenswrapper[4959]: I1007 14:00:14.524934 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/444c0979-8132-4868-bd91-5df3278b7660-secret-volume\") pod \"444c0979-8132-4868-bd91-5df3278b7660\" (UID: \"444c0979-8132-4868-bd91-5df3278b7660\") " Oct 07 14:00:14 crc kubenswrapper[4959]: I1007 14:00:14.525825 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/444c0979-8132-4868-bd91-5df3278b7660-config-volume" (OuterVolumeSpecName: "config-volume") pod "444c0979-8132-4868-bd91-5df3278b7660" (UID: "444c0979-8132-4868-bd91-5df3278b7660"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:00:14 crc kubenswrapper[4959]: I1007 14:00:14.526294 4959 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/444c0979-8132-4868-bd91-5df3278b7660-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 14:00:14 crc kubenswrapper[4959]: I1007 14:00:14.530004 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/444c0979-8132-4868-bd91-5df3278b7660-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "444c0979-8132-4868-bd91-5df3278b7660" (UID: "444c0979-8132-4868-bd91-5df3278b7660"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:00:14 crc kubenswrapper[4959]: I1007 14:00:14.536262 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/444c0979-8132-4868-bd91-5df3278b7660-kube-api-access-fthz5" (OuterVolumeSpecName: "kube-api-access-fthz5") pod "444c0979-8132-4868-bd91-5df3278b7660" (UID: "444c0979-8132-4868-bd91-5df3278b7660"). InnerVolumeSpecName "kube-api-access-fthz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:00:14 crc kubenswrapper[4959]: I1007 14:00:14.627730 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fthz5\" (UniqueName: \"kubernetes.io/projected/444c0979-8132-4868-bd91-5df3278b7660-kube-api-access-fthz5\") on node \"crc\" DevicePath \"\"" Oct 07 14:00:14 crc kubenswrapper[4959]: I1007 14:00:14.627787 4959 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/444c0979-8132-4868-bd91-5df3278b7660-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 07 14:00:15 crc kubenswrapper[4959]: I1007 14:00:15.069242 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330760-wn777" Oct 07 14:00:15 crc kubenswrapper[4959]: I1007 14:00:15.069275 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330760-wn777" event={"ID":"444c0979-8132-4868-bd91-5df3278b7660","Type":"ContainerDied","Data":"0c9fff9615e8c94ccbd13e072b777f0327eb2fdd47b46efdd8703bca8013539f"} Oct 07 14:00:15 crc kubenswrapper[4959]: I1007 14:00:15.069346 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c9fff9615e8c94ccbd13e072b777f0327eb2fdd47b46efdd8703bca8013539f" Oct 07 14:00:15 crc kubenswrapper[4959]: I1007 14:00:15.073070 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-p6mgb" Oct 07 14:00:15 crc kubenswrapper[4959]: I1007 14:00:15.426790 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-85vgl" Oct 07 14:00:15 crc kubenswrapper[4959]: I1007 14:00:15.480917 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-pnbwt" Oct 07 14:00:15 crc kubenswrapper[4959]: I1007 14:00:15.693095 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-hhbqh" Oct 07 14:00:15 crc kubenswrapper[4959]: I1007 14:00:15.774013 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-fd648f65-bc7rh" Oct 07 14:00:15 crc kubenswrapper[4959]: I1007 14:00:15.841590 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-7x6df" Oct 07 14:00:16 crc kubenswrapper[4959]: I1007 14:00:16.031218 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-k4pkf" Oct 07 14:00:16 crc kubenswrapper[4959]: I1007 14:00:16.050740 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-mgrc2" Oct 07 14:00:16 crc kubenswrapper[4959]: I1007 14:00:16.102690 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-nms79" Oct 07 14:00:16 crc kubenswrapper[4959]: I1007 14:00:16.169659 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-rdj2q" Oct 07 14:00:16 crc kubenswrapper[4959]: I1007 14:00:16.183248 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-74gv9" Oct 07 14:00:16 crc kubenswrapper[4959]: I1007 14:00:16.217509 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-76d5577b-4tp8k" Oct 07 14:00:17 crc kubenswrapper[4959]: I1007 14:00:17.336293 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9" Oct 07 14:00:20 crc kubenswrapper[4959]: I1007 14:00:20.629819 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:00:20 crc kubenswrapper[4959]: I1007 14:00:20.630929 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:00:25 crc kubenswrapper[4959]: I1007 14:00:25.415369 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-9k7x7" Oct 07 14:00:25 crc kubenswrapper[4959]: I1007 14:00:25.646352 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-l4hdq" Oct 07 14:00:25 crc kubenswrapper[4959]: I1007 14:00:25.682175 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-k5rmd" Oct 07 14:00:25 crc kubenswrapper[4959]: I1007 14:00:25.714722 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-ksjkq" Oct 07 14:00:25 crc kubenswrapper[4959]: I1007 14:00:25.788825 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-q56bw" Oct 07 14:00:25 crc kubenswrapper[4959]: I1007 14:00:25.825296 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-tfrz7" Oct 07 14:00:25 crc kubenswrapper[4959]: I1007 14:00:25.939318 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-zdxb9" Oct 07 14:00:27 crc kubenswrapper[4959]: I1007 14:00:27.163089 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-lcjnz" event={"ID":"793bfbb8-e13c-4c1c-a490-d422b79b88c5","Type":"ContainerStarted","Data":"c899899e67e8be9039e08b3700ddcffc56765ae33022b0ad5ee45c55439a628d"} Oct 07 14:00:27 crc kubenswrapper[4959]: I1007 14:00:27.163757 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-lcjnz" Oct 07 14:00:27 crc kubenswrapper[4959]: I1007 14:00:27.183078 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-lcjnz" podStartSLOduration=2.630950467 podStartE2EDuration="42.183056382s" podCreationTimestamp="2025-10-07 13:59:45 +0000 UTC" firstStartedPulling="2025-10-07 13:59:47.342803765 +0000 UTC m=+869.426208090" lastFinishedPulling="2025-10-07 14:00:26.89490968 +0000 UTC m=+908.978314005" observedRunningTime="2025-10-07 14:00:27.181850642 +0000 UTC m=+909.265255007" watchObservedRunningTime="2025-10-07 14:00:27.183056382 +0000 UTC m=+909.266460707" Oct 07 14:00:35 crc kubenswrapper[4959]: I1007 14:00:35.980298 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-lcjnz" Oct 07 14:00:50 crc kubenswrapper[4959]: I1007 14:00:50.630164 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:00:50 crc kubenswrapper[4959]: I1007 14:00:50.630732 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:00:54 crc kubenswrapper[4959]: I1007 14:00:54.941053 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-zfb5d"] Oct 07 14:00:54 crc kubenswrapper[4959]: E1007 14:00:54.941615 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="444c0979-8132-4868-bd91-5df3278b7660" containerName="collect-profiles" Oct 07 14:00:54 crc kubenswrapper[4959]: I1007 14:00:54.941632 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="444c0979-8132-4868-bd91-5df3278b7660" containerName="collect-profiles" Oct 07 14:00:54 crc kubenswrapper[4959]: I1007 14:00:54.941792 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="444c0979-8132-4868-bd91-5df3278b7660" containerName="collect-profiles" Oct 07 14:00:54 crc kubenswrapper[4959]: I1007 14:00:54.942803 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-zfb5d" Oct 07 14:00:54 crc kubenswrapper[4959]: I1007 14:00:54.944767 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-j67bk" Oct 07 14:00:54 crc kubenswrapper[4959]: I1007 14:00:54.945494 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 07 14:00:54 crc kubenswrapper[4959]: I1007 14:00:54.945505 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 07 14:00:54 crc kubenswrapper[4959]: I1007 14:00:54.945575 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 07 14:00:54 crc kubenswrapper[4959]: I1007 14:00:54.954477 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-zfb5d"] Oct 07 14:00:54 crc kubenswrapper[4959]: I1007 14:00:54.995074 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-qdx25"] Oct 07 14:00:54 crc kubenswrapper[4959]: I1007 14:00:54.996296 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-qdx25" Oct 07 14:00:54 crc kubenswrapper[4959]: I1007 14:00:54.998518 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 07 14:00:55 crc kubenswrapper[4959]: I1007 14:00:55.009442 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-qdx25"] Oct 07 14:00:55 crc kubenswrapper[4959]: I1007 14:00:55.133252 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3bdddda-a729-4e84-a838-bbae789fee1c-config\") pod \"dnsmasq-dns-758b79db4c-qdx25\" (UID: \"d3bdddda-a729-4e84-a838-bbae789fee1c\") " pod="openstack/dnsmasq-dns-758b79db4c-qdx25" Oct 07 14:00:55 crc kubenswrapper[4959]: I1007 14:00:55.133941 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3bdddda-a729-4e84-a838-bbae789fee1c-dns-svc\") pod \"dnsmasq-dns-758b79db4c-qdx25\" (UID: \"d3bdddda-a729-4e84-a838-bbae789fee1c\") " pod="openstack/dnsmasq-dns-758b79db4c-qdx25" Oct 07 14:00:55 crc kubenswrapper[4959]: I1007 14:00:55.134217 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49b12bae-5f7f-4c6e-a926-06dc9547164c-config\") pod \"dnsmasq-dns-7bfcb9d745-zfb5d\" (UID: \"49b12bae-5f7f-4c6e-a926-06dc9547164c\") " pod="openstack/dnsmasq-dns-7bfcb9d745-zfb5d" Oct 07 14:00:55 crc kubenswrapper[4959]: I1007 14:00:55.134281 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zchqq\" (UniqueName: \"kubernetes.io/projected/49b12bae-5f7f-4c6e-a926-06dc9547164c-kube-api-access-zchqq\") pod \"dnsmasq-dns-7bfcb9d745-zfb5d\" (UID: \"49b12bae-5f7f-4c6e-a926-06dc9547164c\") " pod="openstack/dnsmasq-dns-7bfcb9d745-zfb5d" Oct 07 14:00:55 crc kubenswrapper[4959]: I1007 14:00:55.134311 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s67zk\" (UniqueName: \"kubernetes.io/projected/d3bdddda-a729-4e84-a838-bbae789fee1c-kube-api-access-s67zk\") pod \"dnsmasq-dns-758b79db4c-qdx25\" (UID: \"d3bdddda-a729-4e84-a838-bbae789fee1c\") " pod="openstack/dnsmasq-dns-758b79db4c-qdx25" Oct 07 14:00:55 crc kubenswrapper[4959]: I1007 14:00:55.236258 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3bdddda-a729-4e84-a838-bbae789fee1c-dns-svc\") pod \"dnsmasq-dns-758b79db4c-qdx25\" (UID: \"d3bdddda-a729-4e84-a838-bbae789fee1c\") " pod="openstack/dnsmasq-dns-758b79db4c-qdx25" Oct 07 14:00:55 crc kubenswrapper[4959]: I1007 14:00:55.236303 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3bdddda-a729-4e84-a838-bbae789fee1c-config\") pod \"dnsmasq-dns-758b79db4c-qdx25\" (UID: \"d3bdddda-a729-4e84-a838-bbae789fee1c\") " pod="openstack/dnsmasq-dns-758b79db4c-qdx25" Oct 07 14:00:55 crc kubenswrapper[4959]: I1007 14:00:55.236365 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49b12bae-5f7f-4c6e-a926-06dc9547164c-config\") pod \"dnsmasq-dns-7bfcb9d745-zfb5d\" (UID: \"49b12bae-5f7f-4c6e-a926-06dc9547164c\") " pod="openstack/dnsmasq-dns-7bfcb9d745-zfb5d" Oct 07 14:00:55 crc kubenswrapper[4959]: I1007 14:00:55.236390 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zchqq\" (UniqueName: \"kubernetes.io/projected/49b12bae-5f7f-4c6e-a926-06dc9547164c-kube-api-access-zchqq\") pod \"dnsmasq-dns-7bfcb9d745-zfb5d\" (UID: \"49b12bae-5f7f-4c6e-a926-06dc9547164c\") " pod="openstack/dnsmasq-dns-7bfcb9d745-zfb5d" Oct 07 14:00:55 crc kubenswrapper[4959]: I1007 14:00:55.236410 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s67zk\" (UniqueName: \"kubernetes.io/projected/d3bdddda-a729-4e84-a838-bbae789fee1c-kube-api-access-s67zk\") pod \"dnsmasq-dns-758b79db4c-qdx25\" (UID: \"d3bdddda-a729-4e84-a838-bbae789fee1c\") " pod="openstack/dnsmasq-dns-758b79db4c-qdx25" Oct 07 14:00:55 crc kubenswrapper[4959]: I1007 14:00:55.237225 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3bdddda-a729-4e84-a838-bbae789fee1c-dns-svc\") pod \"dnsmasq-dns-758b79db4c-qdx25\" (UID: \"d3bdddda-a729-4e84-a838-bbae789fee1c\") " pod="openstack/dnsmasq-dns-758b79db4c-qdx25" Oct 07 14:00:55 crc kubenswrapper[4959]: I1007 14:00:55.237263 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3bdddda-a729-4e84-a838-bbae789fee1c-config\") pod \"dnsmasq-dns-758b79db4c-qdx25\" (UID: \"d3bdddda-a729-4e84-a838-bbae789fee1c\") " pod="openstack/dnsmasq-dns-758b79db4c-qdx25" Oct 07 14:00:55 crc kubenswrapper[4959]: I1007 14:00:55.237501 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49b12bae-5f7f-4c6e-a926-06dc9547164c-config\") pod \"dnsmasq-dns-7bfcb9d745-zfb5d\" (UID: \"49b12bae-5f7f-4c6e-a926-06dc9547164c\") " pod="openstack/dnsmasq-dns-7bfcb9d745-zfb5d" Oct 07 14:00:55 crc kubenswrapper[4959]: I1007 14:00:55.256130 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zchqq\" (UniqueName: \"kubernetes.io/projected/49b12bae-5f7f-4c6e-a926-06dc9547164c-kube-api-access-zchqq\") pod \"dnsmasq-dns-7bfcb9d745-zfb5d\" (UID: \"49b12bae-5f7f-4c6e-a926-06dc9547164c\") " pod="openstack/dnsmasq-dns-7bfcb9d745-zfb5d" Oct 07 14:00:55 crc kubenswrapper[4959]: I1007 14:00:55.256434 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s67zk\" (UniqueName: \"kubernetes.io/projected/d3bdddda-a729-4e84-a838-bbae789fee1c-kube-api-access-s67zk\") pod \"dnsmasq-dns-758b79db4c-qdx25\" (UID: \"d3bdddda-a729-4e84-a838-bbae789fee1c\") " pod="openstack/dnsmasq-dns-758b79db4c-qdx25" Oct 07 14:00:55 crc kubenswrapper[4959]: I1007 14:00:55.265336 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-zfb5d" Oct 07 14:00:55 crc kubenswrapper[4959]: I1007 14:00:55.319256 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-qdx25" Oct 07 14:00:55 crc kubenswrapper[4959]: I1007 14:00:55.622539 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-qdx25"] Oct 07 14:00:55 crc kubenswrapper[4959]: I1007 14:00:55.634766 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 14:00:55 crc kubenswrapper[4959]: I1007 14:00:55.733978 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-zfb5d"] Oct 07 14:00:56 crc kubenswrapper[4959]: I1007 14:00:56.444535 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-758b79db4c-qdx25" event={"ID":"d3bdddda-a729-4e84-a838-bbae789fee1c","Type":"ContainerStarted","Data":"d808236e4bb17167cb1a5f4887400cd570321ef7d27992634b8c2c1957017f78"} Oct 07 14:00:56 crc kubenswrapper[4959]: I1007 14:00:56.445479 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bfcb9d745-zfb5d" event={"ID":"49b12bae-5f7f-4c6e-a926-06dc9547164c","Type":"ContainerStarted","Data":"c37e3df83e8e9fcce2443f406f99d0df553ac50c0e96b6b83345cb5747bdebb1"} Oct 07 14:00:56 crc kubenswrapper[4959]: I1007 14:00:56.788448 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-qdx25"] Oct 07 14:00:56 crc kubenswrapper[4959]: I1007 14:00:56.822648 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8575fc99d7-s9zj7"] Oct 07 14:00:56 crc kubenswrapper[4959]: I1007 14:00:56.833910 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8575fc99d7-s9zj7" Oct 07 14:00:56 crc kubenswrapper[4959]: I1007 14:00:56.841896 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8575fc99d7-s9zj7"] Oct 07 14:00:56 crc kubenswrapper[4959]: I1007 14:00:56.859971 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5tjt\" (UniqueName: \"kubernetes.io/projected/2b85003b-b72d-45c0-9657-7e479b192655-kube-api-access-v5tjt\") pod \"dnsmasq-dns-8575fc99d7-s9zj7\" (UID: \"2b85003b-b72d-45c0-9657-7e479b192655\") " pod="openstack/dnsmasq-dns-8575fc99d7-s9zj7" Oct 07 14:00:56 crc kubenswrapper[4959]: I1007 14:00:56.860113 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b85003b-b72d-45c0-9657-7e479b192655-config\") pod \"dnsmasq-dns-8575fc99d7-s9zj7\" (UID: \"2b85003b-b72d-45c0-9657-7e479b192655\") " pod="openstack/dnsmasq-dns-8575fc99d7-s9zj7" Oct 07 14:00:56 crc kubenswrapper[4959]: I1007 14:00:56.860361 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b85003b-b72d-45c0-9657-7e479b192655-dns-svc\") pod \"dnsmasq-dns-8575fc99d7-s9zj7\" (UID: \"2b85003b-b72d-45c0-9657-7e479b192655\") " pod="openstack/dnsmasq-dns-8575fc99d7-s9zj7" Oct 07 14:00:56 crc kubenswrapper[4959]: I1007 14:00:56.961816 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b85003b-b72d-45c0-9657-7e479b192655-dns-svc\") pod \"dnsmasq-dns-8575fc99d7-s9zj7\" (UID: \"2b85003b-b72d-45c0-9657-7e479b192655\") " pod="openstack/dnsmasq-dns-8575fc99d7-s9zj7" Oct 07 14:00:56 crc kubenswrapper[4959]: I1007 14:00:56.962158 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5tjt\" (UniqueName: \"kubernetes.io/projected/2b85003b-b72d-45c0-9657-7e479b192655-kube-api-access-v5tjt\") pod \"dnsmasq-dns-8575fc99d7-s9zj7\" (UID: \"2b85003b-b72d-45c0-9657-7e479b192655\") " pod="openstack/dnsmasq-dns-8575fc99d7-s9zj7" Oct 07 14:00:56 crc kubenswrapper[4959]: I1007 14:00:56.962264 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b85003b-b72d-45c0-9657-7e479b192655-config\") pod \"dnsmasq-dns-8575fc99d7-s9zj7\" (UID: \"2b85003b-b72d-45c0-9657-7e479b192655\") " pod="openstack/dnsmasq-dns-8575fc99d7-s9zj7" Oct 07 14:00:56 crc kubenswrapper[4959]: I1007 14:00:56.962800 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b85003b-b72d-45c0-9657-7e479b192655-dns-svc\") pod \"dnsmasq-dns-8575fc99d7-s9zj7\" (UID: \"2b85003b-b72d-45c0-9657-7e479b192655\") " pod="openstack/dnsmasq-dns-8575fc99d7-s9zj7" Oct 07 14:00:56 crc kubenswrapper[4959]: I1007 14:00:56.963074 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b85003b-b72d-45c0-9657-7e479b192655-config\") pod \"dnsmasq-dns-8575fc99d7-s9zj7\" (UID: \"2b85003b-b72d-45c0-9657-7e479b192655\") " pod="openstack/dnsmasq-dns-8575fc99d7-s9zj7" Oct 07 14:00:56 crc kubenswrapper[4959]: I1007 14:00:56.990366 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5tjt\" (UniqueName: \"kubernetes.io/projected/2b85003b-b72d-45c0-9657-7e479b192655-kube-api-access-v5tjt\") pod \"dnsmasq-dns-8575fc99d7-s9zj7\" (UID: \"2b85003b-b72d-45c0-9657-7e479b192655\") " pod="openstack/dnsmasq-dns-8575fc99d7-s9zj7" Oct 07 14:00:57 crc kubenswrapper[4959]: I1007 14:00:57.094308 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-zfb5d"] Oct 07 14:00:57 crc kubenswrapper[4959]: I1007 14:00:57.119158 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77597f887-m74r4"] Oct 07 14:00:57 crc kubenswrapper[4959]: I1007 14:00:57.120257 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-m74r4" Oct 07 14:00:57 crc kubenswrapper[4959]: I1007 14:00:57.137554 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77597f887-m74r4"] Oct 07 14:00:57 crc kubenswrapper[4959]: I1007 14:00:57.164153 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8575fc99d7-s9zj7" Oct 07 14:00:57 crc kubenswrapper[4959]: I1007 14:00:57.165491 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfptd\" (UniqueName: \"kubernetes.io/projected/08e6890c-fbed-4862-aade-762ee2784a6a-kube-api-access-kfptd\") pod \"dnsmasq-dns-77597f887-m74r4\" (UID: \"08e6890c-fbed-4862-aade-762ee2784a6a\") " pod="openstack/dnsmasq-dns-77597f887-m74r4" Oct 07 14:00:57 crc kubenswrapper[4959]: I1007 14:00:57.165518 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08e6890c-fbed-4862-aade-762ee2784a6a-config\") pod \"dnsmasq-dns-77597f887-m74r4\" (UID: \"08e6890c-fbed-4862-aade-762ee2784a6a\") " pod="openstack/dnsmasq-dns-77597f887-m74r4" Oct 07 14:00:57 crc kubenswrapper[4959]: I1007 14:00:57.165555 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08e6890c-fbed-4862-aade-762ee2784a6a-dns-svc\") pod \"dnsmasq-dns-77597f887-m74r4\" (UID: \"08e6890c-fbed-4862-aade-762ee2784a6a\") " pod="openstack/dnsmasq-dns-77597f887-m74r4" Oct 07 14:00:57 crc kubenswrapper[4959]: I1007 14:00:57.266550 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08e6890c-fbed-4862-aade-762ee2784a6a-dns-svc\") pod \"dnsmasq-dns-77597f887-m74r4\" (UID: \"08e6890c-fbed-4862-aade-762ee2784a6a\") " pod="openstack/dnsmasq-dns-77597f887-m74r4" Oct 07 14:00:57 crc kubenswrapper[4959]: I1007 14:00:57.266663 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfptd\" (UniqueName: \"kubernetes.io/projected/08e6890c-fbed-4862-aade-762ee2784a6a-kube-api-access-kfptd\") pod \"dnsmasq-dns-77597f887-m74r4\" (UID: \"08e6890c-fbed-4862-aade-762ee2784a6a\") " pod="openstack/dnsmasq-dns-77597f887-m74r4" Oct 07 14:00:57 crc kubenswrapper[4959]: I1007 14:00:57.266685 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08e6890c-fbed-4862-aade-762ee2784a6a-config\") pod \"dnsmasq-dns-77597f887-m74r4\" (UID: \"08e6890c-fbed-4862-aade-762ee2784a6a\") " pod="openstack/dnsmasq-dns-77597f887-m74r4" Oct 07 14:00:57 crc kubenswrapper[4959]: I1007 14:00:57.267880 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08e6890c-fbed-4862-aade-762ee2784a6a-config\") pod \"dnsmasq-dns-77597f887-m74r4\" (UID: \"08e6890c-fbed-4862-aade-762ee2784a6a\") " pod="openstack/dnsmasq-dns-77597f887-m74r4" Oct 07 14:00:57 crc kubenswrapper[4959]: I1007 14:00:57.268005 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08e6890c-fbed-4862-aade-762ee2784a6a-dns-svc\") pod \"dnsmasq-dns-77597f887-m74r4\" (UID: \"08e6890c-fbed-4862-aade-762ee2784a6a\") " pod="openstack/dnsmasq-dns-77597f887-m74r4" Oct 07 14:00:57 crc kubenswrapper[4959]: I1007 14:00:57.304180 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfptd\" (UniqueName: \"kubernetes.io/projected/08e6890c-fbed-4862-aade-762ee2784a6a-kube-api-access-kfptd\") pod \"dnsmasq-dns-77597f887-m74r4\" (UID: \"08e6890c-fbed-4862-aade-762ee2784a6a\") " pod="openstack/dnsmasq-dns-77597f887-m74r4" Oct 07 14:00:57 crc kubenswrapper[4959]: I1007 14:00:57.437280 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-m74r4" Oct 07 14:00:57 crc kubenswrapper[4959]: I1007 14:00:57.672977 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8575fc99d7-s9zj7"] Oct 07 14:00:57 crc kubenswrapper[4959]: I1007 14:00:57.995203 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 14:00:57 crc kubenswrapper[4959]: I1007 14:00:57.996682 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.003195 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.003263 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.003195 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.003634 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-xr4q8" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.003830 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.004026 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.004294 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.014142 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.025298 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77597f887-m74r4"] Oct 07 14:00:58 crc kubenswrapper[4959]: W1007 14:00:58.035765 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08e6890c_fbed_4862_aade_762ee2784a6a.slice/crio-8d4e714e8215866b1d6eaf5c5858270da969873396864ef36238450b19370e81 WatchSource:0}: Error finding container 8d4e714e8215866b1d6eaf5c5858270da969873396864ef36238450b19370e81: Status 404 returned error can't find the container with id 8d4e714e8215866b1d6eaf5c5858270da969873396864ef36238450b19370e81 Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.178363 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbdmf\" (UniqueName: \"kubernetes.io/projected/e206a4fa-c178-4f85-be59-46edf7e78833-kube-api-access-fbdmf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.178401 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e206a4fa-c178-4f85-be59-46edf7e78833-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.178442 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e206a4fa-c178-4f85-be59-46edf7e78833-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.178465 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e206a4fa-c178-4f85-be59-46edf7e78833-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.178505 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.178520 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e206a4fa-c178-4f85-be59-46edf7e78833-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.178535 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e206a4fa-c178-4f85-be59-46edf7e78833-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.178572 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e206a4fa-c178-4f85-be59-46edf7e78833-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.178599 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e206a4fa-c178-4f85-be59-46edf7e78833-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.178623 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e206a4fa-c178-4f85-be59-46edf7e78833-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.178641 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e206a4fa-c178-4f85-be59-46edf7e78833-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.268974 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.270301 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.274543 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.274701 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.275022 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.275143 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-8jwt2" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.277226 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.277659 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.278512 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.281901 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbdmf\" (UniqueName: \"kubernetes.io/projected/e206a4fa-c178-4f85-be59-46edf7e78833-kube-api-access-fbdmf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.282524 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e206a4fa-c178-4f85-be59-46edf7e78833-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.282576 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e206a4fa-c178-4f85-be59-46edf7e78833-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.282613 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e206a4fa-c178-4f85-be59-46edf7e78833-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.282671 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.282696 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e206a4fa-c178-4f85-be59-46edf7e78833-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.282723 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e206a4fa-c178-4f85-be59-46edf7e78833-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.282750 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e206a4fa-c178-4f85-be59-46edf7e78833-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.282783 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e206a4fa-c178-4f85-be59-46edf7e78833-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.282817 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e206a4fa-c178-4f85-be59-46edf7e78833-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.282843 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e206a4fa-c178-4f85-be59-46edf7e78833-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.284038 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e206a4fa-c178-4f85-be59-46edf7e78833-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.292385 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.293379 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.294690 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e206a4fa-c178-4f85-be59-46edf7e78833-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.300641 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e206a4fa-c178-4f85-be59-46edf7e78833-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.300971 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e206a4fa-c178-4f85-be59-46edf7e78833-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.301321 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e206a4fa-c178-4f85-be59-46edf7e78833-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.301903 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e206a4fa-c178-4f85-be59-46edf7e78833-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.302244 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e206a4fa-c178-4f85-be59-46edf7e78833-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.307078 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbdmf\" (UniqueName: \"kubernetes.io/projected/e206a4fa-c178-4f85-be59-46edf7e78833-kube-api-access-fbdmf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.308268 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e206a4fa-c178-4f85-be59-46edf7e78833-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.317433 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e206a4fa-c178-4f85-be59-46edf7e78833-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.326234 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.386779 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/90af7ce4-adf2-4722-a726-cc4d7d29056f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.386822 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/90af7ce4-adf2-4722-a726-cc4d7d29056f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.386859 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/90af7ce4-adf2-4722-a726-cc4d7d29056f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.386880 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.386900 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/90af7ce4-adf2-4722-a726-cc4d7d29056f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.387171 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/90af7ce4-adf2-4722-a726-cc4d7d29056f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.387322 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tw6fp\" (UniqueName: \"kubernetes.io/projected/90af7ce4-adf2-4722-a726-cc4d7d29056f-kube-api-access-tw6fp\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.387350 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/90af7ce4-adf2-4722-a726-cc4d7d29056f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.387378 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/90af7ce4-adf2-4722-a726-cc4d7d29056f-config-data\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.387522 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/90af7ce4-adf2-4722-a726-cc4d7d29056f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.387548 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/90af7ce4-adf2-4722-a726-cc4d7d29056f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.470124 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8575fc99d7-s9zj7" event={"ID":"2b85003b-b72d-45c0-9657-7e479b192655","Type":"ContainerStarted","Data":"3d2f573fbf21f5768852bcd04dc0e3f9d90aa0dbc8fc994a75da5e5b3c0d4762"} Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.473175 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77597f887-m74r4" event={"ID":"08e6890c-fbed-4862-aade-762ee2784a6a","Type":"ContainerStarted","Data":"8d4e714e8215866b1d6eaf5c5858270da969873396864ef36238450b19370e81"} Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.489180 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.489233 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/90af7ce4-adf2-4722-a726-cc4d7d29056f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.489287 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/90af7ce4-adf2-4722-a726-cc4d7d29056f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.489329 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/90af7ce4-adf2-4722-a726-cc4d7d29056f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.489353 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tw6fp\" (UniqueName: \"kubernetes.io/projected/90af7ce4-adf2-4722-a726-cc4d7d29056f-kube-api-access-tw6fp\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.489380 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/90af7ce4-adf2-4722-a726-cc4d7d29056f-config-data\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.489417 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/90af7ce4-adf2-4722-a726-cc4d7d29056f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.489437 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/90af7ce4-adf2-4722-a726-cc4d7d29056f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.489473 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/90af7ce4-adf2-4722-a726-cc4d7d29056f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.489495 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/90af7ce4-adf2-4722-a726-cc4d7d29056f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.489542 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/90af7ce4-adf2-4722-a726-cc4d7d29056f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.489539 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.490638 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/90af7ce4-adf2-4722-a726-cc4d7d29056f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.490682 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/90af7ce4-adf2-4722-a726-cc4d7d29056f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.490711 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/90af7ce4-adf2-4722-a726-cc4d7d29056f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.491630 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/90af7ce4-adf2-4722-a726-cc4d7d29056f-config-data\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.492548 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/90af7ce4-adf2-4722-a726-cc4d7d29056f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.494127 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/90af7ce4-adf2-4722-a726-cc4d7d29056f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.494172 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/90af7ce4-adf2-4722-a726-cc4d7d29056f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.496252 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/90af7ce4-adf2-4722-a726-cc4d7d29056f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.505422 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tw6fp\" (UniqueName: \"kubernetes.io/projected/90af7ce4-adf2-4722-a726-cc4d7d29056f-kube-api-access-tw6fp\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.509636 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/90af7ce4-adf2-4722-a726-cc4d7d29056f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.549955 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " pod="openstack/rabbitmq-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.627606 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:00:58 crc kubenswrapper[4959]: I1007 14:00:58.661760 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 07 14:00:59 crc kubenswrapper[4959]: I1007 14:00:59.166037 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 14:00:59 crc kubenswrapper[4959]: W1007 14:00:59.177030 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode206a4fa_c178_4f85_be59_46edf7e78833.slice/crio-4c72abc157662698c847fb5ffdb71559f582932a858fb0d4e6cbe9c90d872ed4 WatchSource:0}: Error finding container 4c72abc157662698c847fb5ffdb71559f582932a858fb0d4e6cbe9c90d872ed4: Status 404 returned error can't find the container with id 4c72abc157662698c847fb5ffdb71559f582932a858fb0d4e6cbe9c90d872ed4 Oct 07 14:00:59 crc kubenswrapper[4959]: I1007 14:00:59.194564 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 14:00:59 crc kubenswrapper[4959]: W1007 14:00:59.215039 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90af7ce4_adf2_4722_a726_cc4d7d29056f.slice/crio-d27b48c1a514c89af66f80c59ec74ff8ac32d2fe06105b31c679db28fabf3a52 WatchSource:0}: Error finding container d27b48c1a514c89af66f80c59ec74ff8ac32d2fe06105b31c679db28fabf3a52: Status 404 returned error can't find the container with id d27b48c1a514c89af66f80c59ec74ff8ac32d2fe06105b31c679db28fabf3a52 Oct 07 14:00:59 crc kubenswrapper[4959]: I1007 14:00:59.486416 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"90af7ce4-adf2-4722-a726-cc4d7d29056f","Type":"ContainerStarted","Data":"d27b48c1a514c89af66f80c59ec74ff8ac32d2fe06105b31c679db28fabf3a52"} Oct 07 14:00:59 crc kubenswrapper[4959]: I1007 14:00:59.488054 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e206a4fa-c178-4f85-be59-46edf7e78833","Type":"ContainerStarted","Data":"4c72abc157662698c847fb5ffdb71559f582932a858fb0d4e6cbe9c90d872ed4"} Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.577030 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.578628 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.592238 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-fp7ms" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.592323 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.592431 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.592438 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.592462 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.598791 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.599063 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.723946 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb9134cb-b4a5-42d6-b0ca-b4d3e1174625-operator-scripts\") pod \"openstack-galera-0\" (UID: \"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625\") " pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.724008 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/fb9134cb-b4a5-42d6-b0ca-b4d3e1174625-secrets\") pod \"openstack-galera-0\" (UID: \"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625\") " pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.724029 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fb9134cb-b4a5-42d6-b0ca-b4d3e1174625-kolla-config\") pod \"openstack-galera-0\" (UID: \"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625\") " pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.724134 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb9134cb-b4a5-42d6-b0ca-b4d3e1174625-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625\") " pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.724162 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625\") " pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.724192 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9134cb-b4a5-42d6-b0ca-b4d3e1174625-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625\") " pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.724207 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq7jg\" (UniqueName: \"kubernetes.io/projected/fb9134cb-b4a5-42d6-b0ca-b4d3e1174625-kube-api-access-kq7jg\") pod \"openstack-galera-0\" (UID: \"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625\") " pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.724227 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fb9134cb-b4a5-42d6-b0ca-b4d3e1174625-config-data-default\") pod \"openstack-galera-0\" (UID: \"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625\") " pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.724247 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fb9134cb-b4a5-42d6-b0ca-b4d3e1174625-config-data-generated\") pod \"openstack-galera-0\" (UID: \"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625\") " pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.825278 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb9134cb-b4a5-42d6-b0ca-b4d3e1174625-operator-scripts\") pod \"openstack-galera-0\" (UID: \"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625\") " pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.825322 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/fb9134cb-b4a5-42d6-b0ca-b4d3e1174625-secrets\") pod \"openstack-galera-0\" (UID: \"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625\") " pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.825341 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fb9134cb-b4a5-42d6-b0ca-b4d3e1174625-kolla-config\") pod \"openstack-galera-0\" (UID: \"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625\") " pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.825415 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb9134cb-b4a5-42d6-b0ca-b4d3e1174625-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625\") " pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.825438 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625\") " pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.825497 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9134cb-b4a5-42d6-b0ca-b4d3e1174625-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625\") " pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.825518 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq7jg\" (UniqueName: \"kubernetes.io/projected/fb9134cb-b4a5-42d6-b0ca-b4d3e1174625-kube-api-access-kq7jg\") pod \"openstack-galera-0\" (UID: \"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625\") " pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.825541 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fb9134cb-b4a5-42d6-b0ca-b4d3e1174625-config-data-default\") pod \"openstack-galera-0\" (UID: \"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625\") " pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.825579 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fb9134cb-b4a5-42d6-b0ca-b4d3e1174625-config-data-generated\") pod \"openstack-galera-0\" (UID: \"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625\") " pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.826057 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fb9134cb-b4a5-42d6-b0ca-b4d3e1174625-config-data-generated\") pod \"openstack-galera-0\" (UID: \"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625\") " pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.828533 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb9134cb-b4a5-42d6-b0ca-b4d3e1174625-operator-scripts\") pod \"openstack-galera-0\" (UID: \"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625\") " pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.831818 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.832680 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fb9134cb-b4a5-42d6-b0ca-b4d3e1174625-config-data-default\") pod \"openstack-galera-0\" (UID: \"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625\") " pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.838374 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fb9134cb-b4a5-42d6-b0ca-b4d3e1174625-kolla-config\") pod \"openstack-galera-0\" (UID: \"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625\") " pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.838428 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9134cb-b4a5-42d6-b0ca-b4d3e1174625-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625\") " pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.841116 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/fb9134cb-b4a5-42d6-b0ca-b4d3e1174625-secrets\") pod \"openstack-galera-0\" (UID: \"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625\") " pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.841480 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb9134cb-b4a5-42d6-b0ca-b4d3e1174625-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625\") " pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.847890 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq7jg\" (UniqueName: \"kubernetes.io/projected/fb9134cb-b4a5-42d6-b0ca-b4d3e1174625-kube-api-access-kq7jg\") pod \"openstack-galera-0\" (UID: \"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625\") " pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.858859 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625\") " pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.917548 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.977094 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.978511 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.980652 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.980693 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-7cvlh" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.980927 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.980982 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 07 14:01:00 crc kubenswrapper[4959]: I1007 14:01:00.988711 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.132391 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6af5bf69-0961-4d34-9cc6-3b04eb48bca1-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"6af5bf69-0961-4d34-9cc6-3b04eb48bca1\") " pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.132782 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"6af5bf69-0961-4d34-9cc6-3b04eb48bca1\") " pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.132865 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j9k4\" (UniqueName: \"kubernetes.io/projected/6af5bf69-0961-4d34-9cc6-3b04eb48bca1-kube-api-access-2j9k4\") pod \"openstack-cell1-galera-0\" (UID: \"6af5bf69-0961-4d34-9cc6-3b04eb48bca1\") " pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.132887 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6af5bf69-0961-4d34-9cc6-3b04eb48bca1-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"6af5bf69-0961-4d34-9cc6-3b04eb48bca1\") " pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.133076 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6af5bf69-0961-4d34-9cc6-3b04eb48bca1-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"6af5bf69-0961-4d34-9cc6-3b04eb48bca1\") " pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.133161 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6af5bf69-0961-4d34-9cc6-3b04eb48bca1-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"6af5bf69-0961-4d34-9cc6-3b04eb48bca1\") " pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.133200 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6af5bf69-0961-4d34-9cc6-3b04eb48bca1-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"6af5bf69-0961-4d34-9cc6-3b04eb48bca1\") " pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.133289 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6af5bf69-0961-4d34-9cc6-3b04eb48bca1-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"6af5bf69-0961-4d34-9cc6-3b04eb48bca1\") " pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.133340 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/6af5bf69-0961-4d34-9cc6-3b04eb48bca1-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"6af5bf69-0961-4d34-9cc6-3b04eb48bca1\") " pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.234381 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6af5bf69-0961-4d34-9cc6-3b04eb48bca1-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"6af5bf69-0961-4d34-9cc6-3b04eb48bca1\") " pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.234434 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6af5bf69-0961-4d34-9cc6-3b04eb48bca1-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"6af5bf69-0961-4d34-9cc6-3b04eb48bca1\") " pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.234488 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6af5bf69-0961-4d34-9cc6-3b04eb48bca1-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"6af5bf69-0961-4d34-9cc6-3b04eb48bca1\") " pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.235599 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6af5bf69-0961-4d34-9cc6-3b04eb48bca1-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"6af5bf69-0961-4d34-9cc6-3b04eb48bca1\") " pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.235999 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6af5bf69-0961-4d34-9cc6-3b04eb48bca1-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"6af5bf69-0961-4d34-9cc6-3b04eb48bca1\") " pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.236049 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/6af5bf69-0961-4d34-9cc6-3b04eb48bca1-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"6af5bf69-0961-4d34-9cc6-3b04eb48bca1\") " pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.236125 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6af5bf69-0961-4d34-9cc6-3b04eb48bca1-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"6af5bf69-0961-4d34-9cc6-3b04eb48bca1\") " pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.236153 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"6af5bf69-0961-4d34-9cc6-3b04eb48bca1\") " pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.236207 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j9k4\" (UniqueName: \"kubernetes.io/projected/6af5bf69-0961-4d34-9cc6-3b04eb48bca1-kube-api-access-2j9k4\") pod \"openstack-cell1-galera-0\" (UID: \"6af5bf69-0961-4d34-9cc6-3b04eb48bca1\") " pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.236226 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6af5bf69-0961-4d34-9cc6-3b04eb48bca1-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"6af5bf69-0961-4d34-9cc6-3b04eb48bca1\") " pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.236286 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6af5bf69-0961-4d34-9cc6-3b04eb48bca1-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"6af5bf69-0961-4d34-9cc6-3b04eb48bca1\") " pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.236599 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"6af5bf69-0961-4d34-9cc6-3b04eb48bca1\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.237378 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6af5bf69-0961-4d34-9cc6-3b04eb48bca1-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"6af5bf69-0961-4d34-9cc6-3b04eb48bca1\") " pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.237633 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6af5bf69-0961-4d34-9cc6-3b04eb48bca1-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"6af5bf69-0961-4d34-9cc6-3b04eb48bca1\") " pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.240828 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6af5bf69-0961-4d34-9cc6-3b04eb48bca1-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"6af5bf69-0961-4d34-9cc6-3b04eb48bca1\") " pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.240880 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6af5bf69-0961-4d34-9cc6-3b04eb48bca1-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"6af5bf69-0961-4d34-9cc6-3b04eb48bca1\") " pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.247201 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/6af5bf69-0961-4d34-9cc6-3b04eb48bca1-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"6af5bf69-0961-4d34-9cc6-3b04eb48bca1\") " pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.255510 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j9k4\" (UniqueName: \"kubernetes.io/projected/6af5bf69-0961-4d34-9cc6-3b04eb48bca1-kube-api-access-2j9k4\") pod \"openstack-cell1-galera-0\" (UID: \"6af5bf69-0961-4d34-9cc6-3b04eb48bca1\") " pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.266794 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"6af5bf69-0961-4d34-9cc6-3b04eb48bca1\") " pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.300856 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.554926 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.561926 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.564449 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.564582 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.576522 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-hszs9" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.577247 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.660817 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/938d13f8-ccf7-4723-a7ff-3e16f7571eac-kolla-config\") pod \"memcached-0\" (UID: \"938d13f8-ccf7-4723-a7ff-3e16f7571eac\") " pod="openstack/memcached-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.660864 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/938d13f8-ccf7-4723-a7ff-3e16f7571eac-memcached-tls-certs\") pod \"memcached-0\" (UID: \"938d13f8-ccf7-4723-a7ff-3e16f7571eac\") " pod="openstack/memcached-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.660895 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rndwx\" (UniqueName: \"kubernetes.io/projected/938d13f8-ccf7-4723-a7ff-3e16f7571eac-kube-api-access-rndwx\") pod \"memcached-0\" (UID: \"938d13f8-ccf7-4723-a7ff-3e16f7571eac\") " pod="openstack/memcached-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.660959 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/938d13f8-ccf7-4723-a7ff-3e16f7571eac-combined-ca-bundle\") pod \"memcached-0\" (UID: \"938d13f8-ccf7-4723-a7ff-3e16f7571eac\") " pod="openstack/memcached-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.660979 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/938d13f8-ccf7-4723-a7ff-3e16f7571eac-config-data\") pod \"memcached-0\" (UID: \"938d13f8-ccf7-4723-a7ff-3e16f7571eac\") " pod="openstack/memcached-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.762155 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/938d13f8-ccf7-4723-a7ff-3e16f7571eac-kolla-config\") pod \"memcached-0\" (UID: \"938d13f8-ccf7-4723-a7ff-3e16f7571eac\") " pod="openstack/memcached-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.762201 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/938d13f8-ccf7-4723-a7ff-3e16f7571eac-memcached-tls-certs\") pod \"memcached-0\" (UID: \"938d13f8-ccf7-4723-a7ff-3e16f7571eac\") " pod="openstack/memcached-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.762243 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rndwx\" (UniqueName: \"kubernetes.io/projected/938d13f8-ccf7-4723-a7ff-3e16f7571eac-kube-api-access-rndwx\") pod \"memcached-0\" (UID: \"938d13f8-ccf7-4723-a7ff-3e16f7571eac\") " pod="openstack/memcached-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.762308 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/938d13f8-ccf7-4723-a7ff-3e16f7571eac-combined-ca-bundle\") pod \"memcached-0\" (UID: \"938d13f8-ccf7-4723-a7ff-3e16f7571eac\") " pod="openstack/memcached-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.762342 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/938d13f8-ccf7-4723-a7ff-3e16f7571eac-config-data\") pod \"memcached-0\" (UID: \"938d13f8-ccf7-4723-a7ff-3e16f7571eac\") " pod="openstack/memcached-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.763072 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/938d13f8-ccf7-4723-a7ff-3e16f7571eac-config-data\") pod \"memcached-0\" (UID: \"938d13f8-ccf7-4723-a7ff-3e16f7571eac\") " pod="openstack/memcached-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.763070 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/938d13f8-ccf7-4723-a7ff-3e16f7571eac-kolla-config\") pod \"memcached-0\" (UID: \"938d13f8-ccf7-4723-a7ff-3e16f7571eac\") " pod="openstack/memcached-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.767456 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/938d13f8-ccf7-4723-a7ff-3e16f7571eac-memcached-tls-certs\") pod \"memcached-0\" (UID: \"938d13f8-ccf7-4723-a7ff-3e16f7571eac\") " pod="openstack/memcached-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.768012 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/938d13f8-ccf7-4723-a7ff-3e16f7571eac-combined-ca-bundle\") pod \"memcached-0\" (UID: \"938d13f8-ccf7-4723-a7ff-3e16f7571eac\") " pod="openstack/memcached-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.785758 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rndwx\" (UniqueName: \"kubernetes.io/projected/938d13f8-ccf7-4723-a7ff-3e16f7571eac-kube-api-access-rndwx\") pod \"memcached-0\" (UID: \"938d13f8-ccf7-4723-a7ff-3e16f7571eac\") " pod="openstack/memcached-0" Oct 07 14:01:01 crc kubenswrapper[4959]: I1007 14:01:01.941841 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 07 14:01:03 crc kubenswrapper[4959]: I1007 14:01:03.304620 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 14:01:03 crc kubenswrapper[4959]: I1007 14:01:03.305884 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 07 14:01:03 crc kubenswrapper[4959]: I1007 14:01:03.308577 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-4srwf" Oct 07 14:01:03 crc kubenswrapper[4959]: I1007 14:01:03.325172 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 14:01:03 crc kubenswrapper[4959]: I1007 14:01:03.386837 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jbht\" (UniqueName: \"kubernetes.io/projected/8b37b4cb-8660-456d-8d54-8e9c7fc83c51-kube-api-access-2jbht\") pod \"kube-state-metrics-0\" (UID: \"8b37b4cb-8660-456d-8d54-8e9c7fc83c51\") " pod="openstack/kube-state-metrics-0" Oct 07 14:01:03 crc kubenswrapper[4959]: I1007 14:01:03.488536 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jbht\" (UniqueName: \"kubernetes.io/projected/8b37b4cb-8660-456d-8d54-8e9c7fc83c51-kube-api-access-2jbht\") pod \"kube-state-metrics-0\" (UID: \"8b37b4cb-8660-456d-8d54-8e9c7fc83c51\") " pod="openstack/kube-state-metrics-0" Oct 07 14:01:03 crc kubenswrapper[4959]: I1007 14:01:03.509536 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jbht\" (UniqueName: \"kubernetes.io/projected/8b37b4cb-8660-456d-8d54-8e9c7fc83c51-kube-api-access-2jbht\") pod \"kube-state-metrics-0\" (UID: \"8b37b4cb-8660-456d-8d54-8e9c7fc83c51\") " pod="openstack/kube-state-metrics-0" Oct 07 14:01:03 crc kubenswrapper[4959]: I1007 14:01:03.632928 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 07 14:01:06 crc kubenswrapper[4959]: I1007 14:01:06.933063 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-q5v9v"] Oct 07 14:01:06 crc kubenswrapper[4959]: I1007 14:01:06.934320 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-q5v9v" Oct 07 14:01:06 crc kubenswrapper[4959]: I1007 14:01:06.943289 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-lttwv" Oct 07 14:01:06 crc kubenswrapper[4959]: I1007 14:01:06.943552 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 07 14:01:06 crc kubenswrapper[4959]: I1007 14:01:06.943586 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 07 14:01:06 crc kubenswrapper[4959]: I1007 14:01:06.954597 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-q5v9v"] Oct 07 14:01:06 crc kubenswrapper[4959]: I1007 14:01:06.993924 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-v6cb5"] Oct 07 14:01:06 crc kubenswrapper[4959]: I1007 14:01:06.998506 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-v6cb5" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.002910 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-v6cb5"] Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.056513 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/603a8fc2-2eac-47b7-8e97-cd117033a313-var-run\") pod \"ovn-controller-q5v9v\" (UID: \"603a8fc2-2eac-47b7-8e97-cd117033a313\") " pod="openstack/ovn-controller-q5v9v" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.056568 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2d0933d8-59fe-4293-998c-8a236af482e1-var-lib\") pod \"ovn-controller-ovs-v6cb5\" (UID: \"2d0933d8-59fe-4293-998c-8a236af482e1\") " pod="openstack/ovn-controller-ovs-v6cb5" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.056592 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2d0933d8-59fe-4293-998c-8a236af482e1-var-run\") pod \"ovn-controller-ovs-v6cb5\" (UID: \"2d0933d8-59fe-4293-998c-8a236af482e1\") " pod="openstack/ovn-controller-ovs-v6cb5" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.056618 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhp67\" (UniqueName: \"kubernetes.io/projected/603a8fc2-2eac-47b7-8e97-cd117033a313-kube-api-access-bhp67\") pod \"ovn-controller-q5v9v\" (UID: \"603a8fc2-2eac-47b7-8e97-cd117033a313\") " pod="openstack/ovn-controller-q5v9v" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.056638 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2d0933d8-59fe-4293-998c-8a236af482e1-var-log\") pod \"ovn-controller-ovs-v6cb5\" (UID: \"2d0933d8-59fe-4293-998c-8a236af482e1\") " pod="openstack/ovn-controller-ovs-v6cb5" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.056669 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/603a8fc2-2eac-47b7-8e97-cd117033a313-scripts\") pod \"ovn-controller-q5v9v\" (UID: \"603a8fc2-2eac-47b7-8e97-cd117033a313\") " pod="openstack/ovn-controller-q5v9v" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.056685 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v555s\" (UniqueName: \"kubernetes.io/projected/2d0933d8-59fe-4293-998c-8a236af482e1-kube-api-access-v555s\") pod \"ovn-controller-ovs-v6cb5\" (UID: \"2d0933d8-59fe-4293-998c-8a236af482e1\") " pod="openstack/ovn-controller-ovs-v6cb5" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.056705 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/603a8fc2-2eac-47b7-8e97-cd117033a313-combined-ca-bundle\") pod \"ovn-controller-q5v9v\" (UID: \"603a8fc2-2eac-47b7-8e97-cd117033a313\") " pod="openstack/ovn-controller-q5v9v" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.057410 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/603a8fc2-2eac-47b7-8e97-cd117033a313-var-log-ovn\") pod \"ovn-controller-q5v9v\" (UID: \"603a8fc2-2eac-47b7-8e97-cd117033a313\") " pod="openstack/ovn-controller-q5v9v" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.057458 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2d0933d8-59fe-4293-998c-8a236af482e1-etc-ovs\") pod \"ovn-controller-ovs-v6cb5\" (UID: \"2d0933d8-59fe-4293-998c-8a236af482e1\") " pod="openstack/ovn-controller-ovs-v6cb5" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.057487 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d0933d8-59fe-4293-998c-8a236af482e1-scripts\") pod \"ovn-controller-ovs-v6cb5\" (UID: \"2d0933d8-59fe-4293-998c-8a236af482e1\") " pod="openstack/ovn-controller-ovs-v6cb5" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.057531 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/603a8fc2-2eac-47b7-8e97-cd117033a313-ovn-controller-tls-certs\") pod \"ovn-controller-q5v9v\" (UID: \"603a8fc2-2eac-47b7-8e97-cd117033a313\") " pod="openstack/ovn-controller-q5v9v" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.058123 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/603a8fc2-2eac-47b7-8e97-cd117033a313-var-run-ovn\") pod \"ovn-controller-q5v9v\" (UID: \"603a8fc2-2eac-47b7-8e97-cd117033a313\") " pod="openstack/ovn-controller-q5v9v" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.159182 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v555s\" (UniqueName: \"kubernetes.io/projected/2d0933d8-59fe-4293-998c-8a236af482e1-kube-api-access-v555s\") pod \"ovn-controller-ovs-v6cb5\" (UID: \"2d0933d8-59fe-4293-998c-8a236af482e1\") " pod="openstack/ovn-controller-ovs-v6cb5" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.159236 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/603a8fc2-2eac-47b7-8e97-cd117033a313-combined-ca-bundle\") pod \"ovn-controller-q5v9v\" (UID: \"603a8fc2-2eac-47b7-8e97-cd117033a313\") " pod="openstack/ovn-controller-q5v9v" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.159286 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/603a8fc2-2eac-47b7-8e97-cd117033a313-var-log-ovn\") pod \"ovn-controller-q5v9v\" (UID: \"603a8fc2-2eac-47b7-8e97-cd117033a313\") " pod="openstack/ovn-controller-q5v9v" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.159303 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2d0933d8-59fe-4293-998c-8a236af482e1-etc-ovs\") pod \"ovn-controller-ovs-v6cb5\" (UID: \"2d0933d8-59fe-4293-998c-8a236af482e1\") " pod="openstack/ovn-controller-ovs-v6cb5" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.159324 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d0933d8-59fe-4293-998c-8a236af482e1-scripts\") pod \"ovn-controller-ovs-v6cb5\" (UID: \"2d0933d8-59fe-4293-998c-8a236af482e1\") " pod="openstack/ovn-controller-ovs-v6cb5" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.159343 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/603a8fc2-2eac-47b7-8e97-cd117033a313-ovn-controller-tls-certs\") pod \"ovn-controller-q5v9v\" (UID: \"603a8fc2-2eac-47b7-8e97-cd117033a313\") " pod="openstack/ovn-controller-q5v9v" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.159362 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/603a8fc2-2eac-47b7-8e97-cd117033a313-var-run-ovn\") pod \"ovn-controller-q5v9v\" (UID: \"603a8fc2-2eac-47b7-8e97-cd117033a313\") " pod="openstack/ovn-controller-q5v9v" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.159403 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/603a8fc2-2eac-47b7-8e97-cd117033a313-var-run\") pod \"ovn-controller-q5v9v\" (UID: \"603a8fc2-2eac-47b7-8e97-cd117033a313\") " pod="openstack/ovn-controller-q5v9v" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.159434 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2d0933d8-59fe-4293-998c-8a236af482e1-var-lib\") pod \"ovn-controller-ovs-v6cb5\" (UID: \"2d0933d8-59fe-4293-998c-8a236af482e1\") " pod="openstack/ovn-controller-ovs-v6cb5" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.159451 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2d0933d8-59fe-4293-998c-8a236af482e1-var-run\") pod \"ovn-controller-ovs-v6cb5\" (UID: \"2d0933d8-59fe-4293-998c-8a236af482e1\") " pod="openstack/ovn-controller-ovs-v6cb5" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.159477 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhp67\" (UniqueName: \"kubernetes.io/projected/603a8fc2-2eac-47b7-8e97-cd117033a313-kube-api-access-bhp67\") pod \"ovn-controller-q5v9v\" (UID: \"603a8fc2-2eac-47b7-8e97-cd117033a313\") " pod="openstack/ovn-controller-q5v9v" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.159493 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2d0933d8-59fe-4293-998c-8a236af482e1-var-log\") pod \"ovn-controller-ovs-v6cb5\" (UID: \"2d0933d8-59fe-4293-998c-8a236af482e1\") " pod="openstack/ovn-controller-ovs-v6cb5" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.159523 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/603a8fc2-2eac-47b7-8e97-cd117033a313-scripts\") pod \"ovn-controller-q5v9v\" (UID: \"603a8fc2-2eac-47b7-8e97-cd117033a313\") " pod="openstack/ovn-controller-q5v9v" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.160000 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/603a8fc2-2eac-47b7-8e97-cd117033a313-var-log-ovn\") pod \"ovn-controller-q5v9v\" (UID: \"603a8fc2-2eac-47b7-8e97-cd117033a313\") " pod="openstack/ovn-controller-q5v9v" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.160156 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2d0933d8-59fe-4293-998c-8a236af482e1-var-run\") pod \"ovn-controller-ovs-v6cb5\" (UID: \"2d0933d8-59fe-4293-998c-8a236af482e1\") " pod="openstack/ovn-controller-ovs-v6cb5" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.160151 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2d0933d8-59fe-4293-998c-8a236af482e1-etc-ovs\") pod \"ovn-controller-ovs-v6cb5\" (UID: \"2d0933d8-59fe-4293-998c-8a236af482e1\") " pod="openstack/ovn-controller-ovs-v6cb5" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.160184 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/603a8fc2-2eac-47b7-8e97-cd117033a313-var-run\") pod \"ovn-controller-q5v9v\" (UID: \"603a8fc2-2eac-47b7-8e97-cd117033a313\") " pod="openstack/ovn-controller-q5v9v" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.160263 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2d0933d8-59fe-4293-998c-8a236af482e1-var-lib\") pod \"ovn-controller-ovs-v6cb5\" (UID: \"2d0933d8-59fe-4293-998c-8a236af482e1\") " pod="openstack/ovn-controller-ovs-v6cb5" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.160344 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2d0933d8-59fe-4293-998c-8a236af482e1-var-log\") pod \"ovn-controller-ovs-v6cb5\" (UID: \"2d0933d8-59fe-4293-998c-8a236af482e1\") " pod="openstack/ovn-controller-ovs-v6cb5" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.160346 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/603a8fc2-2eac-47b7-8e97-cd117033a313-var-run-ovn\") pod \"ovn-controller-q5v9v\" (UID: \"603a8fc2-2eac-47b7-8e97-cd117033a313\") " pod="openstack/ovn-controller-q5v9v" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.163846 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d0933d8-59fe-4293-998c-8a236af482e1-scripts\") pod \"ovn-controller-ovs-v6cb5\" (UID: \"2d0933d8-59fe-4293-998c-8a236af482e1\") " pod="openstack/ovn-controller-ovs-v6cb5" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.164030 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/603a8fc2-2eac-47b7-8e97-cd117033a313-scripts\") pod \"ovn-controller-q5v9v\" (UID: \"603a8fc2-2eac-47b7-8e97-cd117033a313\") " pod="openstack/ovn-controller-q5v9v" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.165753 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/603a8fc2-2eac-47b7-8e97-cd117033a313-ovn-controller-tls-certs\") pod \"ovn-controller-q5v9v\" (UID: \"603a8fc2-2eac-47b7-8e97-cd117033a313\") " pod="openstack/ovn-controller-q5v9v" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.166203 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/603a8fc2-2eac-47b7-8e97-cd117033a313-combined-ca-bundle\") pod \"ovn-controller-q5v9v\" (UID: \"603a8fc2-2eac-47b7-8e97-cd117033a313\") " pod="openstack/ovn-controller-q5v9v" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.179613 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhp67\" (UniqueName: \"kubernetes.io/projected/603a8fc2-2eac-47b7-8e97-cd117033a313-kube-api-access-bhp67\") pod \"ovn-controller-q5v9v\" (UID: \"603a8fc2-2eac-47b7-8e97-cd117033a313\") " pod="openstack/ovn-controller-q5v9v" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.184338 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v555s\" (UniqueName: \"kubernetes.io/projected/2d0933d8-59fe-4293-998c-8a236af482e1-kube-api-access-v555s\") pod \"ovn-controller-ovs-v6cb5\" (UID: \"2d0933d8-59fe-4293-998c-8a236af482e1\") " pod="openstack/ovn-controller-ovs-v6cb5" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.250868 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-q5v9v" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.315025 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-v6cb5" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.834875 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.837072 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.839566 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.840440 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.841061 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-6tplc" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.844270 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.852545 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.860209 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.972243 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/454a13c2-ada4-4885-89b2-efdde4705730-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"454a13c2-ada4-4885-89b2-efdde4705730\") " pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.972311 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/454a13c2-ada4-4885-89b2-efdde4705730-config\") pod \"ovsdbserver-sb-0\" (UID: \"454a13c2-ada4-4885-89b2-efdde4705730\") " pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.972365 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/454a13c2-ada4-4885-89b2-efdde4705730-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"454a13c2-ada4-4885-89b2-efdde4705730\") " pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.972401 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/454a13c2-ada4-4885-89b2-efdde4705730-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"454a13c2-ada4-4885-89b2-efdde4705730\") " pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.972457 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/454a13c2-ada4-4885-89b2-efdde4705730-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"454a13c2-ada4-4885-89b2-efdde4705730\") " pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.972521 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"454a13c2-ada4-4885-89b2-efdde4705730\") " pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.972561 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4rmw\" (UniqueName: \"kubernetes.io/projected/454a13c2-ada4-4885-89b2-efdde4705730-kube-api-access-x4rmw\") pod \"ovsdbserver-sb-0\" (UID: \"454a13c2-ada4-4885-89b2-efdde4705730\") " pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:07 crc kubenswrapper[4959]: I1007 14:01:07.972635 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/454a13c2-ada4-4885-89b2-efdde4705730-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"454a13c2-ada4-4885-89b2-efdde4705730\") " pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:08 crc kubenswrapper[4959]: I1007 14:01:08.074094 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/454a13c2-ada4-4885-89b2-efdde4705730-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"454a13c2-ada4-4885-89b2-efdde4705730\") " pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:08 crc kubenswrapper[4959]: I1007 14:01:08.074184 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/454a13c2-ada4-4885-89b2-efdde4705730-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"454a13c2-ada4-4885-89b2-efdde4705730\") " pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:08 crc kubenswrapper[4959]: I1007 14:01:08.074214 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/454a13c2-ada4-4885-89b2-efdde4705730-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"454a13c2-ada4-4885-89b2-efdde4705730\") " pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:08 crc kubenswrapper[4959]: I1007 14:01:08.074231 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/454a13c2-ada4-4885-89b2-efdde4705730-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"454a13c2-ada4-4885-89b2-efdde4705730\") " pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:08 crc kubenswrapper[4959]: I1007 14:01:08.074247 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/454a13c2-ada4-4885-89b2-efdde4705730-config\") pod \"ovsdbserver-sb-0\" (UID: \"454a13c2-ada4-4885-89b2-efdde4705730\") " pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:08 crc kubenswrapper[4959]: I1007 14:01:08.074280 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/454a13c2-ada4-4885-89b2-efdde4705730-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"454a13c2-ada4-4885-89b2-efdde4705730\") " pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:08 crc kubenswrapper[4959]: I1007 14:01:08.074298 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"454a13c2-ada4-4885-89b2-efdde4705730\") " pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:08 crc kubenswrapper[4959]: I1007 14:01:08.074322 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4rmw\" (UniqueName: \"kubernetes.io/projected/454a13c2-ada4-4885-89b2-efdde4705730-kube-api-access-x4rmw\") pod \"ovsdbserver-sb-0\" (UID: \"454a13c2-ada4-4885-89b2-efdde4705730\") " pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:08 crc kubenswrapper[4959]: I1007 14:01:08.074648 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"454a13c2-ada4-4885-89b2-efdde4705730\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:08 crc kubenswrapper[4959]: I1007 14:01:08.074918 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/454a13c2-ada4-4885-89b2-efdde4705730-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"454a13c2-ada4-4885-89b2-efdde4705730\") " pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:08 crc kubenswrapper[4959]: I1007 14:01:08.075276 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/454a13c2-ada4-4885-89b2-efdde4705730-config\") pod \"ovsdbserver-sb-0\" (UID: \"454a13c2-ada4-4885-89b2-efdde4705730\") " pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:08 crc kubenswrapper[4959]: I1007 14:01:08.076275 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/454a13c2-ada4-4885-89b2-efdde4705730-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"454a13c2-ada4-4885-89b2-efdde4705730\") " pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:08 crc kubenswrapper[4959]: I1007 14:01:08.078213 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/454a13c2-ada4-4885-89b2-efdde4705730-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"454a13c2-ada4-4885-89b2-efdde4705730\") " pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:08 crc kubenswrapper[4959]: I1007 14:01:08.079443 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/454a13c2-ada4-4885-89b2-efdde4705730-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"454a13c2-ada4-4885-89b2-efdde4705730\") " pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:08 crc kubenswrapper[4959]: I1007 14:01:08.080697 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/454a13c2-ada4-4885-89b2-efdde4705730-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"454a13c2-ada4-4885-89b2-efdde4705730\") " pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:08 crc kubenswrapper[4959]: I1007 14:01:08.093756 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4rmw\" (UniqueName: \"kubernetes.io/projected/454a13c2-ada4-4885-89b2-efdde4705730-kube-api-access-x4rmw\") pod \"ovsdbserver-sb-0\" (UID: \"454a13c2-ada4-4885-89b2-efdde4705730\") " pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:08 crc kubenswrapper[4959]: I1007 14:01:08.095661 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"454a13c2-ada4-4885-89b2-efdde4705730\") " pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:08 crc kubenswrapper[4959]: I1007 14:01:08.209657 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:09 crc kubenswrapper[4959]: I1007 14:01:09.958841 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 07 14:01:09 crc kubenswrapper[4959]: I1007 14:01:09.961667 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:09 crc kubenswrapper[4959]: I1007 14:01:09.966569 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 07 14:01:09 crc kubenswrapper[4959]: I1007 14:01:09.966612 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-8t4rn" Oct 07 14:01:09 crc kubenswrapper[4959]: I1007 14:01:09.967188 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 07 14:01:09 crc kubenswrapper[4959]: I1007 14:01:09.967622 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 07 14:01:09 crc kubenswrapper[4959]: I1007 14:01:09.972424 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 07 14:01:10 crc kubenswrapper[4959]: I1007 14:01:10.010445 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac9cc42f-0fad-48ad-b4b2-3494a5d10a61-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ac9cc42f-0fad-48ad-b4b2-3494a5d10a61\") " pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:10 crc kubenswrapper[4959]: I1007 14:01:10.010489 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ac9cc42f-0fad-48ad-b4b2-3494a5d10a61\") " pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:10 crc kubenswrapper[4959]: I1007 14:01:10.010513 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac9cc42f-0fad-48ad-b4b2-3494a5d10a61-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ac9cc42f-0fad-48ad-b4b2-3494a5d10a61\") " pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:10 crc kubenswrapper[4959]: I1007 14:01:10.010548 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac9cc42f-0fad-48ad-b4b2-3494a5d10a61-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ac9cc42f-0fad-48ad-b4b2-3494a5d10a61\") " pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:10 crc kubenswrapper[4959]: I1007 14:01:10.010568 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ac9cc42f-0fad-48ad-b4b2-3494a5d10a61-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ac9cc42f-0fad-48ad-b4b2-3494a5d10a61\") " pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:10 crc kubenswrapper[4959]: I1007 14:01:10.010603 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk2c2\" (UniqueName: \"kubernetes.io/projected/ac9cc42f-0fad-48ad-b4b2-3494a5d10a61-kube-api-access-dk2c2\") pod \"ovsdbserver-nb-0\" (UID: \"ac9cc42f-0fad-48ad-b4b2-3494a5d10a61\") " pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:10 crc kubenswrapper[4959]: I1007 14:01:10.010630 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac9cc42f-0fad-48ad-b4b2-3494a5d10a61-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ac9cc42f-0fad-48ad-b4b2-3494a5d10a61\") " pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:10 crc kubenswrapper[4959]: I1007 14:01:10.010653 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac9cc42f-0fad-48ad-b4b2-3494a5d10a61-config\") pod \"ovsdbserver-nb-0\" (UID: \"ac9cc42f-0fad-48ad-b4b2-3494a5d10a61\") " pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:10 crc kubenswrapper[4959]: I1007 14:01:10.113031 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk2c2\" (UniqueName: \"kubernetes.io/projected/ac9cc42f-0fad-48ad-b4b2-3494a5d10a61-kube-api-access-dk2c2\") pod \"ovsdbserver-nb-0\" (UID: \"ac9cc42f-0fad-48ad-b4b2-3494a5d10a61\") " pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:10 crc kubenswrapper[4959]: I1007 14:01:10.113143 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac9cc42f-0fad-48ad-b4b2-3494a5d10a61-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ac9cc42f-0fad-48ad-b4b2-3494a5d10a61\") " pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:10 crc kubenswrapper[4959]: I1007 14:01:10.113528 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac9cc42f-0fad-48ad-b4b2-3494a5d10a61-config\") pod \"ovsdbserver-nb-0\" (UID: \"ac9cc42f-0fad-48ad-b4b2-3494a5d10a61\") " pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:10 crc kubenswrapper[4959]: I1007 14:01:10.113610 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac9cc42f-0fad-48ad-b4b2-3494a5d10a61-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ac9cc42f-0fad-48ad-b4b2-3494a5d10a61\") " pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:10 crc kubenswrapper[4959]: I1007 14:01:10.113653 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ac9cc42f-0fad-48ad-b4b2-3494a5d10a61\") " pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:10 crc kubenswrapper[4959]: I1007 14:01:10.113763 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac9cc42f-0fad-48ad-b4b2-3494a5d10a61-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ac9cc42f-0fad-48ad-b4b2-3494a5d10a61\") " pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:10 crc kubenswrapper[4959]: I1007 14:01:10.114116 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ac9cc42f-0fad-48ad-b4b2-3494a5d10a61\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:10 crc kubenswrapper[4959]: I1007 14:01:10.114870 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac9cc42f-0fad-48ad-b4b2-3494a5d10a61-config\") pod \"ovsdbserver-nb-0\" (UID: \"ac9cc42f-0fad-48ad-b4b2-3494a5d10a61\") " pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:10 crc kubenswrapper[4959]: I1007 14:01:10.115064 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac9cc42f-0fad-48ad-b4b2-3494a5d10a61-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ac9cc42f-0fad-48ad-b4b2-3494a5d10a61\") " pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:10 crc kubenswrapper[4959]: I1007 14:01:10.115148 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac9cc42f-0fad-48ad-b4b2-3494a5d10a61-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ac9cc42f-0fad-48ad-b4b2-3494a5d10a61\") " pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:10 crc kubenswrapper[4959]: I1007 14:01:10.115180 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ac9cc42f-0fad-48ad-b4b2-3494a5d10a61-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ac9cc42f-0fad-48ad-b4b2-3494a5d10a61\") " pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:10 crc kubenswrapper[4959]: I1007 14:01:10.115640 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ac9cc42f-0fad-48ad-b4b2-3494a5d10a61-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ac9cc42f-0fad-48ad-b4b2-3494a5d10a61\") " pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:10 crc kubenswrapper[4959]: I1007 14:01:10.121149 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac9cc42f-0fad-48ad-b4b2-3494a5d10a61-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ac9cc42f-0fad-48ad-b4b2-3494a5d10a61\") " pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:10 crc kubenswrapper[4959]: I1007 14:01:10.122012 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac9cc42f-0fad-48ad-b4b2-3494a5d10a61-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ac9cc42f-0fad-48ad-b4b2-3494a5d10a61\") " pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:10 crc kubenswrapper[4959]: I1007 14:01:10.129201 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac9cc42f-0fad-48ad-b4b2-3494a5d10a61-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ac9cc42f-0fad-48ad-b4b2-3494a5d10a61\") " pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:10 crc kubenswrapper[4959]: I1007 14:01:10.131893 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk2c2\" (UniqueName: \"kubernetes.io/projected/ac9cc42f-0fad-48ad-b4b2-3494a5d10a61-kube-api-access-dk2c2\") pod \"ovsdbserver-nb-0\" (UID: \"ac9cc42f-0fad-48ad-b4b2-3494a5d10a61\") " pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:10 crc kubenswrapper[4959]: I1007 14:01:10.141182 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ac9cc42f-0fad-48ad-b4b2-3494a5d10a61\") " pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:10 crc kubenswrapper[4959]: I1007 14:01:10.290737 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:18 crc kubenswrapper[4959]: I1007 14:01:18.743642 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 07 14:01:19 crc kubenswrapper[4959]: W1007 14:01:19.136181 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod938d13f8_ccf7_4723_a7ff_3e16f7571eac.slice/crio-5d552bdb2d7c34126ec3c4c7197942527a7085bf3c2e27a0f7ea0a7e13b8394f WatchSource:0}: Error finding container 5d552bdb2d7c34126ec3c4c7197942527a7085bf3c2e27a0f7ea0a7e13b8394f: Status 404 returned error can't find the container with id 5d552bdb2d7c34126ec3c4c7197942527a7085bf3c2e27a0f7ea0a7e13b8394f Oct 07 14:01:19 crc kubenswrapper[4959]: E1007 14:01:19.169451 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df" Oct 07 14:01:19 crc kubenswrapper[4959]: E1007 14:01:19.169627 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kfptd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-77597f887-m74r4_openstack(08e6890c-fbed-4862-aade-762ee2784a6a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 14:01:19 crc kubenswrapper[4959]: E1007 14:01:19.171082 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-77597f887-m74r4" podUID="08e6890c-fbed-4862-aade-762ee2784a6a" Oct 07 14:01:19 crc kubenswrapper[4959]: E1007 14:01:19.200597 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df" Oct 07 14:01:19 crc kubenswrapper[4959]: E1007 14:01:19.200868 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zchqq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7bfcb9d745-zfb5d_openstack(49b12bae-5f7f-4c6e-a926-06dc9547164c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 14:01:19 crc kubenswrapper[4959]: E1007 14:01:19.202223 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7bfcb9d745-zfb5d" podUID="49b12bae-5f7f-4c6e-a926-06dc9547164c" Oct 07 14:01:19 crc kubenswrapper[4959]: E1007 14:01:19.207299 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df" Oct 07 14:01:19 crc kubenswrapper[4959]: E1007 14:01:19.207682 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s67zk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-758b79db4c-qdx25_openstack(d3bdddda-a729-4e84-a838-bbae789fee1c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 14:01:19 crc kubenswrapper[4959]: E1007 14:01:19.208946 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-758b79db4c-qdx25" podUID="d3bdddda-a729-4e84-a838-bbae789fee1c" Oct 07 14:01:19 crc kubenswrapper[4959]: E1007 14:01:19.233079 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df" Oct 07 14:01:19 crc kubenswrapper[4959]: E1007 14:01:19.233425 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v5tjt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-8575fc99d7-s9zj7_openstack(2b85003b-b72d-45c0-9657-7e479b192655): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 14:01:19 crc kubenswrapper[4959]: E1007 14:01:19.234723 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-8575fc99d7-s9zj7" podUID="2b85003b-b72d-45c0-9657-7e479b192655" Oct 07 14:01:19 crc kubenswrapper[4959]: I1007 14:01:19.633542 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 07 14:01:19 crc kubenswrapper[4959]: W1007 14:01:19.641227 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb9134cb_b4a5_42d6_b0ca_b4d3e1174625.slice/crio-639efe8b9620bc25c503734803d3157df2e6caef68f6a614137014db853069dc WatchSource:0}: Error finding container 639efe8b9620bc25c503734803d3157df2e6caef68f6a614137014db853069dc: Status 404 returned error can't find the container with id 639efe8b9620bc25c503734803d3157df2e6caef68f6a614137014db853069dc Oct 07 14:01:19 crc kubenswrapper[4959]: I1007 14:01:19.644708 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"938d13f8-ccf7-4723-a7ff-3e16f7571eac","Type":"ContainerStarted","Data":"5d552bdb2d7c34126ec3c4c7197942527a7085bf3c2e27a0f7ea0a7e13b8394f"} Oct 07 14:01:19 crc kubenswrapper[4959]: I1007 14:01:19.646125 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 07 14:01:19 crc kubenswrapper[4959]: E1007 14:01:19.646342 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df\\\"\"" pod="openstack/dnsmasq-dns-8575fc99d7-s9zj7" podUID="2b85003b-b72d-45c0-9657-7e479b192655" Oct 07 14:01:19 crc kubenswrapper[4959]: E1007 14:01:19.646380 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df\\\"\"" pod="openstack/dnsmasq-dns-77597f887-m74r4" podUID="08e6890c-fbed-4862-aade-762ee2784a6a" Oct 07 14:01:19 crc kubenswrapper[4959]: I1007 14:01:19.786364 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 14:01:19 crc kubenswrapper[4959]: I1007 14:01:19.794077 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 07 14:01:19 crc kubenswrapper[4959]: I1007 14:01:19.897199 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-q5v9v"] Oct 07 14:01:19 crc kubenswrapper[4959]: I1007 14:01:19.933473 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-v6cb5"] Oct 07 14:01:19 crc kubenswrapper[4959]: W1007 14:01:19.936390 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d0933d8_59fe_4293_998c_8a236af482e1.slice/crio-6ed04fbf066887afabcb7a3292198e8833bb6b54c8834abb911fdb0f31754cc4 WatchSource:0}: Error finding container 6ed04fbf066887afabcb7a3292198e8833bb6b54c8834abb911fdb0f31754cc4: Status 404 returned error can't find the container with id 6ed04fbf066887afabcb7a3292198e8833bb6b54c8834abb911fdb0f31754cc4 Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.019848 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-zfb5d" Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.024016 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-qdx25" Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.058470 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3bdddda-a729-4e84-a838-bbae789fee1c-dns-svc\") pod \"d3bdddda-a729-4e84-a838-bbae789fee1c\" (UID: \"d3bdddda-a729-4e84-a838-bbae789fee1c\") " Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.058547 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s67zk\" (UniqueName: \"kubernetes.io/projected/d3bdddda-a729-4e84-a838-bbae789fee1c-kube-api-access-s67zk\") pod \"d3bdddda-a729-4e84-a838-bbae789fee1c\" (UID: \"d3bdddda-a729-4e84-a838-bbae789fee1c\") " Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.058634 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49b12bae-5f7f-4c6e-a926-06dc9547164c-config\") pod \"49b12bae-5f7f-4c6e-a926-06dc9547164c\" (UID: \"49b12bae-5f7f-4c6e-a926-06dc9547164c\") " Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.058705 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3bdddda-a729-4e84-a838-bbae789fee1c-config\") pod \"d3bdddda-a729-4e84-a838-bbae789fee1c\" (UID: \"d3bdddda-a729-4e84-a838-bbae789fee1c\") " Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.058823 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zchqq\" (UniqueName: \"kubernetes.io/projected/49b12bae-5f7f-4c6e-a926-06dc9547164c-kube-api-access-zchqq\") pod \"49b12bae-5f7f-4c6e-a926-06dc9547164c\" (UID: \"49b12bae-5f7f-4c6e-a926-06dc9547164c\") " Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.059861 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3bdddda-a729-4e84-a838-bbae789fee1c-config" (OuterVolumeSpecName: "config") pod "d3bdddda-a729-4e84-a838-bbae789fee1c" (UID: "d3bdddda-a729-4e84-a838-bbae789fee1c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.059900 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49b12bae-5f7f-4c6e-a926-06dc9547164c-config" (OuterVolumeSpecName: "config") pod "49b12bae-5f7f-4c6e-a926-06dc9547164c" (UID: "49b12bae-5f7f-4c6e-a926-06dc9547164c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.059976 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3bdddda-a729-4e84-a838-bbae789fee1c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d3bdddda-a729-4e84-a838-bbae789fee1c" (UID: "d3bdddda-a729-4e84-a838-bbae789fee1c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.064144 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49b12bae-5f7f-4c6e-a926-06dc9547164c-kube-api-access-zchqq" (OuterVolumeSpecName: "kube-api-access-zchqq") pod "49b12bae-5f7f-4c6e-a926-06dc9547164c" (UID: "49b12bae-5f7f-4c6e-a926-06dc9547164c"). InnerVolumeSpecName "kube-api-access-zchqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.160541 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3bdddda-a729-4e84-a838-bbae789fee1c-config\") on node \"crc\" DevicePath \"\"" Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.160585 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zchqq\" (UniqueName: \"kubernetes.io/projected/49b12bae-5f7f-4c6e-a926-06dc9547164c-kube-api-access-zchqq\") on node \"crc\" DevicePath \"\"" Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.160625 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3bdddda-a729-4e84-a838-bbae789fee1c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.160639 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49b12bae-5f7f-4c6e-a926-06dc9547164c-config\") on node \"crc\" DevicePath \"\"" Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.173261 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.270430 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3bdddda-a729-4e84-a838-bbae789fee1c-kube-api-access-s67zk" (OuterVolumeSpecName: "kube-api-access-s67zk") pod "d3bdddda-a729-4e84-a838-bbae789fee1c" (UID: "d3bdddda-a729-4e84-a838-bbae789fee1c"). InnerVolumeSpecName "kube-api-access-s67zk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:01:20 crc kubenswrapper[4959]: W1007 14:01:20.275620 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod454a13c2_ada4_4885_89b2_efdde4705730.slice/crio-629f4702ea52cd6026100704e88b5d33157b5e5dd2a8f7d1de1885723d6af778 WatchSource:0}: Error finding container 629f4702ea52cd6026100704e88b5d33157b5e5dd2a8f7d1de1885723d6af778: Status 404 returned error can't find the container with id 629f4702ea52cd6026100704e88b5d33157b5e5dd2a8f7d1de1885723d6af778 Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.363711 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s67zk\" (UniqueName: \"kubernetes.io/projected/d3bdddda-a729-4e84-a838-bbae789fee1c-kube-api-access-s67zk\") on node \"crc\" DevicePath \"\"" Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.630551 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.630614 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.630662 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.631296 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"be8cac39c0f211821bf3cae10b6e2f58082165f269df748f76ccd66a0c6a1240"} pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.631374 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" containerID="cri-o://be8cac39c0f211821bf3cae10b6e2f58082165f269df748f76ccd66a0c6a1240" gracePeriod=600 Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.659002 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-qdx25" Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.664241 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625","Type":"ContainerStarted","Data":"639efe8b9620bc25c503734803d3157df2e6caef68f6a614137014db853069dc"} Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.664278 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"6af5bf69-0961-4d34-9cc6-3b04eb48bca1","Type":"ContainerStarted","Data":"0b23be82783a873a98787a40d479d93fa6b56bc42dc529cb446c0815b1e15763"} Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.664289 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8b37b4cb-8660-456d-8d54-8e9c7fc83c51","Type":"ContainerStarted","Data":"70e8864f749c83dccd7480a3686e850af2d76b658c16cceb91cd1dfcd54d94e7"} Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.664299 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-758b79db4c-qdx25" event={"ID":"d3bdddda-a729-4e84-a838-bbae789fee1c","Type":"ContainerDied","Data":"d808236e4bb17167cb1a5f4887400cd570321ef7d27992634b8c2c1957017f78"} Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.664314 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"454a13c2-ada4-4885-89b2-efdde4705730","Type":"ContainerStarted","Data":"629f4702ea52cd6026100704e88b5d33157b5e5dd2a8f7d1de1885723d6af778"} Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.664810 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e206a4fa-c178-4f85-be59-46edf7e78833","Type":"ContainerStarted","Data":"a9137451d6a4f2897ca96d09e23e311123bffd7256a6e6f4b2714386a1d4c45c"} Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.666171 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"90af7ce4-adf2-4722-a726-cc4d7d29056f","Type":"ContainerStarted","Data":"b6140a1530482e80a1f682355efe01fbc1dd47f08e5fd863d3a4a6fcf5355ee8"} Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.680399 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-q5v9v" event={"ID":"603a8fc2-2eac-47b7-8e97-cd117033a313","Type":"ContainerStarted","Data":"97324182ae9dcd6ea65c620df0a51c1c0d03f00908861576dcdd38b82f1337b3"} Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.702490 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-v6cb5" event={"ID":"2d0933d8-59fe-4293-998c-8a236af482e1","Type":"ContainerStarted","Data":"6ed04fbf066887afabcb7a3292198e8833bb6b54c8834abb911fdb0f31754cc4"} Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.705263 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bfcb9d745-zfb5d" event={"ID":"49b12bae-5f7f-4c6e-a926-06dc9547164c","Type":"ContainerDied","Data":"c37e3df83e8e9fcce2443f406f99d0df553ac50c0e96b6b83345cb5747bdebb1"} Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.705337 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-zfb5d" Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.709298 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ac9cc42f-0fad-48ad-b4b2-3494a5d10a61","Type":"ContainerStarted","Data":"a45aed45a2acb13ef9ef27e137ea228130795ee4b2aee3210b2895ad96cef024"} Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.767762 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-qdx25"] Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.788204 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-qdx25"] Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.801021 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-zfb5d"] Oct 07 14:01:20 crc kubenswrapper[4959]: I1007 14:01:20.806810 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-zfb5d"] Oct 07 14:01:21 crc kubenswrapper[4959]: I1007 14:01:21.733294 4959 generic.go:334] "Generic (PLEG): container finished" podID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerID="be8cac39c0f211821bf3cae10b6e2f58082165f269df748f76ccd66a0c6a1240" exitCode=0 Oct 07 14:01:21 crc kubenswrapper[4959]: I1007 14:01:21.733381 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerDied","Data":"be8cac39c0f211821bf3cae10b6e2f58082165f269df748f76ccd66a0c6a1240"} Oct 07 14:01:21 crc kubenswrapper[4959]: I1007 14:01:21.733473 4959 scope.go:117] "RemoveContainer" containerID="59698f7bbe675b82b270fb68b1462a80d17b9a12f2121c8b8aa53d105e879f24" Oct 07 14:01:22 crc kubenswrapper[4959]: I1007 14:01:22.665755 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49b12bae-5f7f-4c6e-a926-06dc9547164c" path="/var/lib/kubelet/pods/49b12bae-5f7f-4c6e-a926-06dc9547164c/volumes" Oct 07 14:01:22 crc kubenswrapper[4959]: I1007 14:01:22.666211 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3bdddda-a729-4e84-a838-bbae789fee1c" path="/var/lib/kubelet/pods/d3bdddda-a729-4e84-a838-bbae789fee1c/volumes" Oct 07 14:01:23 crc kubenswrapper[4959]: I1007 14:01:23.763877 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerStarted","Data":"f783f953eb3f6009a38800382e284161bac6530f08d9be67139e2b69d6c3e99d"} Oct 07 14:01:26 crc kubenswrapper[4959]: I1007 14:01:26.793909 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"938d13f8-ccf7-4723-a7ff-3e16f7571eac","Type":"ContainerStarted","Data":"294e01e7613dcd0839d34d04aa3b58b484a8841e91046e2000275e1f2237703e"} Oct 07 14:01:26 crc kubenswrapper[4959]: I1007 14:01:26.794458 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 07 14:01:26 crc kubenswrapper[4959]: I1007 14:01:26.822077 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=23.648017103 podStartE2EDuration="25.822042996s" podCreationTimestamp="2025-10-07 14:01:01 +0000 UTC" firstStartedPulling="2025-10-07 14:01:19.150481952 +0000 UTC m=+961.233886277" lastFinishedPulling="2025-10-07 14:01:21.324507845 +0000 UTC m=+963.407912170" observedRunningTime="2025-10-07 14:01:26.819896802 +0000 UTC m=+968.903301157" watchObservedRunningTime="2025-10-07 14:01:26.822042996 +0000 UTC m=+968.905447321" Oct 07 14:01:27 crc kubenswrapper[4959]: I1007 14:01:27.824565 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8b37b4cb-8660-456d-8d54-8e9c7fc83c51","Type":"ContainerStarted","Data":"7691c6913d8e8a52f64be65d70d9957ab25a7302b12a66f302caef479bc76ff1"} Oct 07 14:01:27 crc kubenswrapper[4959]: I1007 14:01:27.825917 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 07 14:01:27 crc kubenswrapper[4959]: I1007 14:01:27.827917 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-q5v9v" event={"ID":"603a8fc2-2eac-47b7-8e97-cd117033a313","Type":"ContainerStarted","Data":"e5bdbc9ef3c8c4a4e6a2ddc87c6fb300dd7efc8bef68fd53da80411699284c8d"} Oct 07 14:01:27 crc kubenswrapper[4959]: I1007 14:01:27.827995 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-q5v9v" Oct 07 14:01:27 crc kubenswrapper[4959]: I1007 14:01:27.830241 4959 generic.go:334] "Generic (PLEG): container finished" podID="2d0933d8-59fe-4293-998c-8a236af482e1" containerID="0abd97f2dc02c26470e7a45e682d428954b8866e847a86443c4373daec4caf03" exitCode=0 Oct 07 14:01:27 crc kubenswrapper[4959]: I1007 14:01:27.830619 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-v6cb5" event={"ID":"2d0933d8-59fe-4293-998c-8a236af482e1","Type":"ContainerDied","Data":"0abd97f2dc02c26470e7a45e682d428954b8866e847a86443c4373daec4caf03"} Oct 07 14:01:27 crc kubenswrapper[4959]: I1007 14:01:27.832260 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"454a13c2-ada4-4885-89b2-efdde4705730","Type":"ContainerStarted","Data":"82cf21612bd46f3dd647a275a18154860d9668c9e658bbb5524d267f162d768b"} Oct 07 14:01:27 crc kubenswrapper[4959]: I1007 14:01:27.833616 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ac9cc42f-0fad-48ad-b4b2-3494a5d10a61","Type":"ContainerStarted","Data":"202467c6987b7a9873d9a06f5e931522b9babd59d4e2ee6474575d9880773d5a"} Oct 07 14:01:27 crc kubenswrapper[4959]: I1007 14:01:27.835364 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625","Type":"ContainerStarted","Data":"0acc1fc599d0360302d286b9afc9eff9eacf500397403e9a74609e16face0a4d"} Oct 07 14:01:27 crc kubenswrapper[4959]: I1007 14:01:27.839522 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"6af5bf69-0961-4d34-9cc6-3b04eb48bca1","Type":"ContainerStarted","Data":"b8305cf50aa9af7fd17daf0ac327fb2813af8666ae6e028a6b2b910d01d81ebc"} Oct 07 14:01:27 crc kubenswrapper[4959]: I1007 14:01:27.845128 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=17.936899734 podStartE2EDuration="24.845087409s" podCreationTimestamp="2025-10-07 14:01:03 +0000 UTC" firstStartedPulling="2025-10-07 14:01:19.798025728 +0000 UTC m=+961.881430053" lastFinishedPulling="2025-10-07 14:01:26.706213413 +0000 UTC m=+968.789617728" observedRunningTime="2025-10-07 14:01:27.840079513 +0000 UTC m=+969.923483838" watchObservedRunningTime="2025-10-07 14:01:27.845087409 +0000 UTC m=+969.928491734" Oct 07 14:01:27 crc kubenswrapper[4959]: I1007 14:01:27.911818 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-q5v9v" podStartSLOduration=15.776963203 podStartE2EDuration="21.911801731s" podCreationTimestamp="2025-10-07 14:01:06 +0000 UTC" firstStartedPulling="2025-10-07 14:01:19.911702116 +0000 UTC m=+961.995106441" lastFinishedPulling="2025-10-07 14:01:26.046540644 +0000 UTC m=+968.129944969" observedRunningTime="2025-10-07 14:01:27.908708363 +0000 UTC m=+969.992112698" watchObservedRunningTime="2025-10-07 14:01:27.911801731 +0000 UTC m=+969.995206056" Oct 07 14:01:28 crc kubenswrapper[4959]: I1007 14:01:28.849249 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-v6cb5" event={"ID":"2d0933d8-59fe-4293-998c-8a236af482e1","Type":"ContainerStarted","Data":"8fb6d4927f9f83e8d54fe316d5df114a4468472f68348e8c646a681f037c33c4"} Oct 07 14:01:28 crc kubenswrapper[4959]: I1007 14:01:28.849570 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-v6cb5" event={"ID":"2d0933d8-59fe-4293-998c-8a236af482e1","Type":"ContainerStarted","Data":"02e2cc6602a981fc7d06e0737c9a01fc9a0b4d5e7d2f7bfa986d3750f850288b"} Oct 07 14:01:28 crc kubenswrapper[4959]: I1007 14:01:28.870802 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-v6cb5" podStartSLOduration=17.073773776 podStartE2EDuration="22.870762269s" podCreationTimestamp="2025-10-07 14:01:06 +0000 UTC" firstStartedPulling="2025-10-07 14:01:19.93862238 +0000 UTC m=+962.022026705" lastFinishedPulling="2025-10-07 14:01:25.735610843 +0000 UTC m=+967.819015198" observedRunningTime="2025-10-07 14:01:28.864890352 +0000 UTC m=+970.948294697" watchObservedRunningTime="2025-10-07 14:01:28.870762269 +0000 UTC m=+970.954166594" Oct 07 14:01:29 crc kubenswrapper[4959]: I1007 14:01:29.855890 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-v6cb5" Oct 07 14:01:29 crc kubenswrapper[4959]: I1007 14:01:29.856218 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-v6cb5" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.168879 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-msjqn"] Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.171004 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-msjqn" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.173225 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.183948 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-msjqn"] Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.237504 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94-combined-ca-bundle\") pod \"ovn-controller-metrics-msjqn\" (UID: \"ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94\") " pod="openstack/ovn-controller-metrics-msjqn" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.237598 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxknd\" (UniqueName: \"kubernetes.io/projected/ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94-kube-api-access-mxknd\") pod \"ovn-controller-metrics-msjqn\" (UID: \"ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94\") " pod="openstack/ovn-controller-metrics-msjqn" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.237726 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94-config\") pod \"ovn-controller-metrics-msjqn\" (UID: \"ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94\") " pod="openstack/ovn-controller-metrics-msjqn" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.237745 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94-ovn-rundir\") pod \"ovn-controller-metrics-msjqn\" (UID: \"ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94\") " pod="openstack/ovn-controller-metrics-msjqn" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.237760 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-msjqn\" (UID: \"ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94\") " pod="openstack/ovn-controller-metrics-msjqn" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.237790 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94-ovs-rundir\") pod \"ovn-controller-metrics-msjqn\" (UID: \"ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94\") " pod="openstack/ovn-controller-metrics-msjqn" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.318487 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8575fc99d7-s9zj7"] Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.339426 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94-config\") pod \"ovn-controller-metrics-msjqn\" (UID: \"ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94\") " pod="openstack/ovn-controller-metrics-msjqn" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.339465 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94-ovn-rundir\") pod \"ovn-controller-metrics-msjqn\" (UID: \"ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94\") " pod="openstack/ovn-controller-metrics-msjqn" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.339483 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-msjqn\" (UID: \"ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94\") " pod="openstack/ovn-controller-metrics-msjqn" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.339516 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94-ovs-rundir\") pod \"ovn-controller-metrics-msjqn\" (UID: \"ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94\") " pod="openstack/ovn-controller-metrics-msjqn" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.339544 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94-combined-ca-bundle\") pod \"ovn-controller-metrics-msjqn\" (UID: \"ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94\") " pod="openstack/ovn-controller-metrics-msjqn" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.339600 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxknd\" (UniqueName: \"kubernetes.io/projected/ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94-kube-api-access-mxknd\") pod \"ovn-controller-metrics-msjqn\" (UID: \"ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94\") " pod="openstack/ovn-controller-metrics-msjqn" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.339865 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94-ovn-rundir\") pod \"ovn-controller-metrics-msjqn\" (UID: \"ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94\") " pod="openstack/ovn-controller-metrics-msjqn" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.340080 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94-ovs-rundir\") pod \"ovn-controller-metrics-msjqn\" (UID: \"ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94\") " pod="openstack/ovn-controller-metrics-msjqn" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.386077 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94-config\") pod \"ovn-controller-metrics-msjqn\" (UID: \"ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94\") " pod="openstack/ovn-controller-metrics-msjqn" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.391410 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-msjqn\" (UID: \"ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94\") " pod="openstack/ovn-controller-metrics-msjqn" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.413249 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94-combined-ca-bundle\") pod \"ovn-controller-metrics-msjqn\" (UID: \"ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94\") " pod="openstack/ovn-controller-metrics-msjqn" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.416626 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxknd\" (UniqueName: \"kubernetes.io/projected/ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94-kube-api-access-mxknd\") pod \"ovn-controller-metrics-msjqn\" (UID: \"ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94\") " pod="openstack/ovn-controller-metrics-msjqn" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.428883 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54c9499b4f-7kc8c"] Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.432093 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54c9499b4f-7kc8c" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.435475 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.438952 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54c9499b4f-7kc8c"] Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.491234 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77597f887-m74r4"] Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.505852 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-msjqn" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.507481 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bc45f6dcf-t6p7q"] Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.512640 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.517575 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.535814 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bc45f6dcf-t6p7q"] Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.543944 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c729d934-aff5-4b38-a0de-aa8526fe1b05-dns-svc\") pod \"dnsmasq-dns-54c9499b4f-7kc8c\" (UID: \"c729d934-aff5-4b38-a0de-aa8526fe1b05\") " pod="openstack/dnsmasq-dns-54c9499b4f-7kc8c" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.544029 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wd85k\" (UniqueName: \"kubernetes.io/projected/c729d934-aff5-4b38-a0de-aa8526fe1b05-kube-api-access-wd85k\") pod \"dnsmasq-dns-54c9499b4f-7kc8c\" (UID: \"c729d934-aff5-4b38-a0de-aa8526fe1b05\") " pod="openstack/dnsmasq-dns-54c9499b4f-7kc8c" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.544249 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c729d934-aff5-4b38-a0de-aa8526fe1b05-config\") pod \"dnsmasq-dns-54c9499b4f-7kc8c\" (UID: \"c729d934-aff5-4b38-a0de-aa8526fe1b05\") " pod="openstack/dnsmasq-dns-54c9499b4f-7kc8c" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.544310 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c729d934-aff5-4b38-a0de-aa8526fe1b05-ovsdbserver-sb\") pod \"dnsmasq-dns-54c9499b4f-7kc8c\" (UID: \"c729d934-aff5-4b38-a0de-aa8526fe1b05\") " pod="openstack/dnsmasq-dns-54c9499b4f-7kc8c" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.645377 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wd85k\" (UniqueName: \"kubernetes.io/projected/c729d934-aff5-4b38-a0de-aa8526fe1b05-kube-api-access-wd85k\") pod \"dnsmasq-dns-54c9499b4f-7kc8c\" (UID: \"c729d934-aff5-4b38-a0de-aa8526fe1b05\") " pod="openstack/dnsmasq-dns-54c9499b4f-7kc8c" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.645827 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c502b70-d657-47d6-a703-4ecf3cc84d01-dns-svc\") pod \"dnsmasq-dns-bc45f6dcf-t6p7q\" (UID: \"4c502b70-d657-47d6-a703-4ecf3cc84d01\") " pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.645896 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwpvl\" (UniqueName: \"kubernetes.io/projected/4c502b70-d657-47d6-a703-4ecf3cc84d01-kube-api-access-hwpvl\") pod \"dnsmasq-dns-bc45f6dcf-t6p7q\" (UID: \"4c502b70-d657-47d6-a703-4ecf3cc84d01\") " pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.645926 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c729d934-aff5-4b38-a0de-aa8526fe1b05-config\") pod \"dnsmasq-dns-54c9499b4f-7kc8c\" (UID: \"c729d934-aff5-4b38-a0de-aa8526fe1b05\") " pod="openstack/dnsmasq-dns-54c9499b4f-7kc8c" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.645962 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c502b70-d657-47d6-a703-4ecf3cc84d01-ovsdbserver-sb\") pod \"dnsmasq-dns-bc45f6dcf-t6p7q\" (UID: \"4c502b70-d657-47d6-a703-4ecf3cc84d01\") " pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.646933 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c729d934-aff5-4b38-a0de-aa8526fe1b05-config\") pod \"dnsmasq-dns-54c9499b4f-7kc8c\" (UID: \"c729d934-aff5-4b38-a0de-aa8526fe1b05\") " pod="openstack/dnsmasq-dns-54c9499b4f-7kc8c" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.646991 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c729d934-aff5-4b38-a0de-aa8526fe1b05-ovsdbserver-sb\") pod \"dnsmasq-dns-54c9499b4f-7kc8c\" (UID: \"c729d934-aff5-4b38-a0de-aa8526fe1b05\") " pod="openstack/dnsmasq-dns-54c9499b4f-7kc8c" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.647058 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c502b70-d657-47d6-a703-4ecf3cc84d01-config\") pod \"dnsmasq-dns-bc45f6dcf-t6p7q\" (UID: \"4c502b70-d657-47d6-a703-4ecf3cc84d01\") " pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.647113 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c502b70-d657-47d6-a703-4ecf3cc84d01-ovsdbserver-nb\") pod \"dnsmasq-dns-bc45f6dcf-t6p7q\" (UID: \"4c502b70-d657-47d6-a703-4ecf3cc84d01\") " pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.647148 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c729d934-aff5-4b38-a0de-aa8526fe1b05-dns-svc\") pod \"dnsmasq-dns-54c9499b4f-7kc8c\" (UID: \"c729d934-aff5-4b38-a0de-aa8526fe1b05\") " pod="openstack/dnsmasq-dns-54c9499b4f-7kc8c" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.647794 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c729d934-aff5-4b38-a0de-aa8526fe1b05-dns-svc\") pod \"dnsmasq-dns-54c9499b4f-7kc8c\" (UID: \"c729d934-aff5-4b38-a0de-aa8526fe1b05\") " pod="openstack/dnsmasq-dns-54c9499b4f-7kc8c" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.648640 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c729d934-aff5-4b38-a0de-aa8526fe1b05-ovsdbserver-sb\") pod \"dnsmasq-dns-54c9499b4f-7kc8c\" (UID: \"c729d934-aff5-4b38-a0de-aa8526fe1b05\") " pod="openstack/dnsmasq-dns-54c9499b4f-7kc8c" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.674958 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wd85k\" (UniqueName: \"kubernetes.io/projected/c729d934-aff5-4b38-a0de-aa8526fe1b05-kube-api-access-wd85k\") pod \"dnsmasq-dns-54c9499b4f-7kc8c\" (UID: \"c729d934-aff5-4b38-a0de-aa8526fe1b05\") " pod="openstack/dnsmasq-dns-54c9499b4f-7kc8c" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.748393 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c502b70-d657-47d6-a703-4ecf3cc84d01-ovsdbserver-nb\") pod \"dnsmasq-dns-bc45f6dcf-t6p7q\" (UID: \"4c502b70-d657-47d6-a703-4ecf3cc84d01\") " pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.748489 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c502b70-d657-47d6-a703-4ecf3cc84d01-dns-svc\") pod \"dnsmasq-dns-bc45f6dcf-t6p7q\" (UID: \"4c502b70-d657-47d6-a703-4ecf3cc84d01\") " pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.748537 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwpvl\" (UniqueName: \"kubernetes.io/projected/4c502b70-d657-47d6-a703-4ecf3cc84d01-kube-api-access-hwpvl\") pod \"dnsmasq-dns-bc45f6dcf-t6p7q\" (UID: \"4c502b70-d657-47d6-a703-4ecf3cc84d01\") " pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.748568 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c502b70-d657-47d6-a703-4ecf3cc84d01-ovsdbserver-sb\") pod \"dnsmasq-dns-bc45f6dcf-t6p7q\" (UID: \"4c502b70-d657-47d6-a703-4ecf3cc84d01\") " pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.748624 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c502b70-d657-47d6-a703-4ecf3cc84d01-config\") pod \"dnsmasq-dns-bc45f6dcf-t6p7q\" (UID: \"4c502b70-d657-47d6-a703-4ecf3cc84d01\") " pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.749570 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c502b70-d657-47d6-a703-4ecf3cc84d01-config\") pod \"dnsmasq-dns-bc45f6dcf-t6p7q\" (UID: \"4c502b70-d657-47d6-a703-4ecf3cc84d01\") " pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.749609 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c502b70-d657-47d6-a703-4ecf3cc84d01-ovsdbserver-nb\") pod \"dnsmasq-dns-bc45f6dcf-t6p7q\" (UID: \"4c502b70-d657-47d6-a703-4ecf3cc84d01\") " pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.749623 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c502b70-d657-47d6-a703-4ecf3cc84d01-ovsdbserver-sb\") pod \"dnsmasq-dns-bc45f6dcf-t6p7q\" (UID: \"4c502b70-d657-47d6-a703-4ecf3cc84d01\") " pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.750007 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c502b70-d657-47d6-a703-4ecf3cc84d01-dns-svc\") pod \"dnsmasq-dns-bc45f6dcf-t6p7q\" (UID: \"4c502b70-d657-47d6-a703-4ecf3cc84d01\") " pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.766293 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwpvl\" (UniqueName: \"kubernetes.io/projected/4c502b70-d657-47d6-a703-4ecf3cc84d01-kube-api-access-hwpvl\") pod \"dnsmasq-dns-bc45f6dcf-t6p7q\" (UID: \"4c502b70-d657-47d6-a703-4ecf3cc84d01\") " pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.806223 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54c9499b4f-7kc8c" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.817751 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8575fc99d7-s9zj7" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.849963 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b85003b-b72d-45c0-9657-7e479b192655-dns-svc\") pod \"2b85003b-b72d-45c0-9657-7e479b192655\" (UID: \"2b85003b-b72d-45c0-9657-7e479b192655\") " Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.850150 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b85003b-b72d-45c0-9657-7e479b192655-config\") pod \"2b85003b-b72d-45c0-9657-7e479b192655\" (UID: \"2b85003b-b72d-45c0-9657-7e479b192655\") " Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.850258 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5tjt\" (UniqueName: \"kubernetes.io/projected/2b85003b-b72d-45c0-9657-7e479b192655-kube-api-access-v5tjt\") pod \"2b85003b-b72d-45c0-9657-7e479b192655\" (UID: \"2b85003b-b72d-45c0-9657-7e479b192655\") " Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.850761 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b85003b-b72d-45c0-9657-7e479b192655-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2b85003b-b72d-45c0-9657-7e479b192655" (UID: "2b85003b-b72d-45c0-9657-7e479b192655"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.850774 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b85003b-b72d-45c0-9657-7e479b192655-config" (OuterVolumeSpecName: "config") pod "2b85003b-b72d-45c0-9657-7e479b192655" (UID: "2b85003b-b72d-45c0-9657-7e479b192655"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.851112 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b85003b-b72d-45c0-9657-7e479b192655-config\") on node \"crc\" DevicePath \"\"" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.851134 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b85003b-b72d-45c0-9657-7e479b192655-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.854067 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b85003b-b72d-45c0-9657-7e479b192655-kube-api-access-v5tjt" (OuterVolumeSpecName: "kube-api-access-v5tjt") pod "2b85003b-b72d-45c0-9657-7e479b192655" (UID: "2b85003b-b72d-45c0-9657-7e479b192655"). InnerVolumeSpecName "kube-api-access-v5tjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.857576 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.864519 4959 generic.go:334] "Generic (PLEG): container finished" podID="fb9134cb-b4a5-42d6-b0ca-b4d3e1174625" containerID="0acc1fc599d0360302d286b9afc9eff9eacf500397403e9a74609e16face0a4d" exitCode=0 Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.864610 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625","Type":"ContainerDied","Data":"0acc1fc599d0360302d286b9afc9eff9eacf500397403e9a74609e16face0a4d"} Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.867431 4959 generic.go:334] "Generic (PLEG): container finished" podID="6af5bf69-0961-4d34-9cc6-3b04eb48bca1" containerID="b8305cf50aa9af7fd17daf0ac327fb2813af8666ae6e028a6b2b910d01d81ebc" exitCode=0 Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.867484 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"6af5bf69-0961-4d34-9cc6-3b04eb48bca1","Type":"ContainerDied","Data":"b8305cf50aa9af7fd17daf0ac327fb2813af8666ae6e028a6b2b910d01d81ebc"} Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.871807 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8575fc99d7-s9zj7" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.871873 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8575fc99d7-s9zj7" event={"ID":"2b85003b-b72d-45c0-9657-7e479b192655","Type":"ContainerDied","Data":"3d2f573fbf21f5768852bcd04dc0e3f9d90aa0dbc8fc994a75da5e5b3c0d4762"} Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.949458 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8575fc99d7-s9zj7"] Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.953147 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5tjt\" (UniqueName: \"kubernetes.io/projected/2b85003b-b72d-45c0-9657-7e479b192655-kube-api-access-v5tjt\") on node \"crc\" DevicePath \"\"" Oct 07 14:01:30 crc kubenswrapper[4959]: I1007 14:01:30.955384 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8575fc99d7-s9zj7"] Oct 07 14:01:31 crc kubenswrapper[4959]: I1007 14:01:31.086727 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-m74r4" Oct 07 14:01:31 crc kubenswrapper[4959]: I1007 14:01:31.156060 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfptd\" (UniqueName: \"kubernetes.io/projected/08e6890c-fbed-4862-aade-762ee2784a6a-kube-api-access-kfptd\") pod \"08e6890c-fbed-4862-aade-762ee2784a6a\" (UID: \"08e6890c-fbed-4862-aade-762ee2784a6a\") " Oct 07 14:01:31 crc kubenswrapper[4959]: I1007 14:01:31.156633 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08e6890c-fbed-4862-aade-762ee2784a6a-config\") pod \"08e6890c-fbed-4862-aade-762ee2784a6a\" (UID: \"08e6890c-fbed-4862-aade-762ee2784a6a\") " Oct 07 14:01:31 crc kubenswrapper[4959]: I1007 14:01:31.156918 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08e6890c-fbed-4862-aade-762ee2784a6a-dns-svc\") pod \"08e6890c-fbed-4862-aade-762ee2784a6a\" (UID: \"08e6890c-fbed-4862-aade-762ee2784a6a\") " Oct 07 14:01:31 crc kubenswrapper[4959]: I1007 14:01:31.157582 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08e6890c-fbed-4862-aade-762ee2784a6a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "08e6890c-fbed-4862-aade-762ee2784a6a" (UID: "08e6890c-fbed-4862-aade-762ee2784a6a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:01:31 crc kubenswrapper[4959]: I1007 14:01:31.157592 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08e6890c-fbed-4862-aade-762ee2784a6a-config" (OuterVolumeSpecName: "config") pod "08e6890c-fbed-4862-aade-762ee2784a6a" (UID: "08e6890c-fbed-4862-aade-762ee2784a6a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:01:31 crc kubenswrapper[4959]: I1007 14:01:31.159495 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08e6890c-fbed-4862-aade-762ee2784a6a-kube-api-access-kfptd" (OuterVolumeSpecName: "kube-api-access-kfptd") pod "08e6890c-fbed-4862-aade-762ee2784a6a" (UID: "08e6890c-fbed-4862-aade-762ee2784a6a"). InnerVolumeSpecName "kube-api-access-kfptd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:01:31 crc kubenswrapper[4959]: I1007 14:01:31.258180 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08e6890c-fbed-4862-aade-762ee2784a6a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 14:01:31 crc kubenswrapper[4959]: I1007 14:01:31.258473 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfptd\" (UniqueName: \"kubernetes.io/projected/08e6890c-fbed-4862-aade-762ee2784a6a-kube-api-access-kfptd\") on node \"crc\" DevicePath \"\"" Oct 07 14:01:31 crc kubenswrapper[4959]: I1007 14:01:31.258484 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08e6890c-fbed-4862-aade-762ee2784a6a-config\") on node \"crc\" DevicePath \"\"" Oct 07 14:01:31 crc kubenswrapper[4959]: I1007 14:01:31.598356 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-msjqn"] Oct 07 14:01:31 crc kubenswrapper[4959]: I1007 14:01:31.603527 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54c9499b4f-7kc8c"] Oct 07 14:01:31 crc kubenswrapper[4959]: I1007 14:01:31.695195 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bc45f6dcf-t6p7q"] Oct 07 14:01:31 crc kubenswrapper[4959]: I1007 14:01:31.878991 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"fb9134cb-b4a5-42d6-b0ca-b4d3e1174625","Type":"ContainerStarted","Data":"4f4e79f7fdafe60131e5f6820b9666bd63692e3eda0d000eecf10771a40af0a4"} Oct 07 14:01:31 crc kubenswrapper[4959]: I1007 14:01:31.884423 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-msjqn" event={"ID":"ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94","Type":"ContainerStarted","Data":"2f91dcd3686bef28d38538a28ac390e704af84d6d5d51a5b7f09b85ded4338f6"} Oct 07 14:01:31 crc kubenswrapper[4959]: I1007 14:01:31.884456 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-msjqn" event={"ID":"ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94","Type":"ContainerStarted","Data":"79fdd50ddb995f00755f94ecb9278d308a0235bb703739dfe5af7bd2bef92f3b"} Oct 07 14:01:31 crc kubenswrapper[4959]: I1007 14:01:31.888149 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77597f887-m74r4" event={"ID":"08e6890c-fbed-4862-aade-762ee2784a6a","Type":"ContainerDied","Data":"8d4e714e8215866b1d6eaf5c5858270da969873396864ef36238450b19370e81"} Oct 07 14:01:31 crc kubenswrapper[4959]: I1007 14:01:31.888197 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-m74r4" Oct 07 14:01:31 crc kubenswrapper[4959]: I1007 14:01:31.901256 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"6af5bf69-0961-4d34-9cc6-3b04eb48bca1","Type":"ContainerStarted","Data":"235f67419635455584e9837ac09311c20ddeec65e3391ef7c47dc3267fa0d43a"} Oct 07 14:01:31 crc kubenswrapper[4959]: I1007 14:01:31.912972 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=26.822167002 podStartE2EDuration="32.912954316s" podCreationTimestamp="2025-10-07 14:00:59 +0000 UTC" firstStartedPulling="2025-10-07 14:01:19.645000173 +0000 UTC m=+961.728404498" lastFinishedPulling="2025-10-07 14:01:25.735787487 +0000 UTC m=+967.819191812" observedRunningTime="2025-10-07 14:01:31.905968501 +0000 UTC m=+973.989372826" watchObservedRunningTime="2025-10-07 14:01:31.912954316 +0000 UTC m=+973.996358651" Oct 07 14:01:31 crc kubenswrapper[4959]: I1007 14:01:31.912978 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"454a13c2-ada4-4885-89b2-efdde4705730","Type":"ContainerStarted","Data":"af8837522adc854e0b35d22082da2abd04c26c0f1b0dad0194b8e6a95803b8e5"} Oct 07 14:01:31 crc kubenswrapper[4959]: I1007 14:01:31.916513 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ac9cc42f-0fad-48ad-b4b2-3494a5d10a61","Type":"ContainerStarted","Data":"343b13e08ec178ceebef7e8a8c70fa10f272b20782bfe4496d7efcf32c76aa8c"} Oct 07 14:01:31 crc kubenswrapper[4959]: I1007 14:01:31.924674 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c9499b4f-7kc8c" event={"ID":"c729d934-aff5-4b38-a0de-aa8526fe1b05","Type":"ContainerStarted","Data":"c26a1e07085b7c468a462dfdbf509d16df7ee5a20e17e49dbb39c7669f27d766"} Oct 07 14:01:31 crc kubenswrapper[4959]: I1007 14:01:31.936303 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" event={"ID":"4c502b70-d657-47d6-a703-4ecf3cc84d01","Type":"ContainerStarted","Data":"742deb80e08dc504344a63df2479689c408d2162fbedec5d05274f611d49dd3e"} Oct 07 14:01:31 crc kubenswrapper[4959]: I1007 14:01:31.946677 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 07 14:01:31 crc kubenswrapper[4959]: I1007 14:01:31.953064 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-msjqn" podStartSLOduration=1.9530459599999999 podStartE2EDuration="1.95304596s" podCreationTimestamp="2025-10-07 14:01:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:01:31.95223811 +0000 UTC m=+974.035642435" watchObservedRunningTime="2025-10-07 14:01:31.95304596 +0000 UTC m=+974.036450285" Oct 07 14:01:31 crc kubenswrapper[4959]: I1007 14:01:31.957593 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=26.92267445 podStartE2EDuration="32.957575844s" podCreationTimestamp="2025-10-07 14:00:59 +0000 UTC" firstStartedPulling="2025-10-07 14:01:19.880949985 +0000 UTC m=+961.964354310" lastFinishedPulling="2025-10-07 14:01:25.915851369 +0000 UTC m=+967.999255704" observedRunningTime="2025-10-07 14:01:31.934867905 +0000 UTC m=+974.018272230" watchObservedRunningTime="2025-10-07 14:01:31.957575844 +0000 UTC m=+974.040980169" Oct 07 14:01:32 crc kubenswrapper[4959]: I1007 14:01:32.014558 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=12.47620837 podStartE2EDuration="24.014537001s" podCreationTimestamp="2025-10-07 14:01:08 +0000 UTC" firstStartedPulling="2025-10-07 14:01:19.644847749 +0000 UTC m=+961.728252064" lastFinishedPulling="2025-10-07 14:01:31.18317637 +0000 UTC m=+973.266580695" observedRunningTime="2025-10-07 14:01:31.997233158 +0000 UTC m=+974.080637483" watchObservedRunningTime="2025-10-07 14:01:32.014537001 +0000 UTC m=+974.097941346" Oct 07 14:01:32 crc kubenswrapper[4959]: I1007 14:01:32.052482 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77597f887-m74r4"] Oct 07 14:01:32 crc kubenswrapper[4959]: I1007 14:01:32.055658 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77597f887-m74r4"] Oct 07 14:01:32 crc kubenswrapper[4959]: I1007 14:01:32.072967 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=15.106814912 podStartE2EDuration="26.072948865s" podCreationTimestamp="2025-10-07 14:01:06 +0000 UTC" firstStartedPulling="2025-10-07 14:01:20.27821734 +0000 UTC m=+962.361621665" lastFinishedPulling="2025-10-07 14:01:31.244351293 +0000 UTC m=+973.327755618" observedRunningTime="2025-10-07 14:01:32.068462442 +0000 UTC m=+974.151866767" watchObservedRunningTime="2025-10-07 14:01:32.072948865 +0000 UTC m=+974.156353190" Oct 07 14:01:32 crc kubenswrapper[4959]: I1007 14:01:32.210143 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:32 crc kubenswrapper[4959]: I1007 14:01:32.251227 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:32 crc kubenswrapper[4959]: I1007 14:01:32.662531 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08e6890c-fbed-4862-aade-762ee2784a6a" path="/var/lib/kubelet/pods/08e6890c-fbed-4862-aade-762ee2784a6a/volumes" Oct 07 14:01:32 crc kubenswrapper[4959]: I1007 14:01:32.663365 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b85003b-b72d-45c0-9657-7e479b192655" path="/var/lib/kubelet/pods/2b85003b-b72d-45c0-9657-7e479b192655/volumes" Oct 07 14:01:32 crc kubenswrapper[4959]: I1007 14:01:32.943426 4959 generic.go:334] "Generic (PLEG): container finished" podID="c729d934-aff5-4b38-a0de-aa8526fe1b05" containerID="ce7d0f7bf1c0c709d5f3cc8e1e688d6ea6647bd5dd8c76d503c55b699afb79c1" exitCode=0 Oct 07 14:01:32 crc kubenswrapper[4959]: I1007 14:01:32.943464 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c9499b4f-7kc8c" event={"ID":"c729d934-aff5-4b38-a0de-aa8526fe1b05","Type":"ContainerDied","Data":"ce7d0f7bf1c0c709d5f3cc8e1e688d6ea6647bd5dd8c76d503c55b699afb79c1"} Oct 07 14:01:32 crc kubenswrapper[4959]: I1007 14:01:32.946613 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" event={"ID":"4c502b70-d657-47d6-a703-4ecf3cc84d01","Type":"ContainerDied","Data":"60b68bdc6665fe51a22286c453a65af6ffea0a6c2901f698c050116f76fbc189"} Oct 07 14:01:32 crc kubenswrapper[4959]: I1007 14:01:32.947146 4959 generic.go:334] "Generic (PLEG): container finished" podID="4c502b70-d657-47d6-a703-4ecf3cc84d01" containerID="60b68bdc6665fe51a22286c453a65af6ffea0a6c2901f698c050116f76fbc189" exitCode=0 Oct 07 14:01:32 crc kubenswrapper[4959]: I1007 14:01:32.947655 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:33 crc kubenswrapper[4959]: I1007 14:01:33.003077 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 07 14:01:33 crc kubenswrapper[4959]: I1007 14:01:33.636606 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 07 14:01:33 crc kubenswrapper[4959]: I1007 14:01:33.955406 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c9499b4f-7kc8c" event={"ID":"c729d934-aff5-4b38-a0de-aa8526fe1b05","Type":"ContainerStarted","Data":"e448092192070ea048ed33d088ebd517a3f52143dcdff42e672bd7a34fcfe546"} Oct 07 14:01:33 crc kubenswrapper[4959]: I1007 14:01:33.955543 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-54c9499b4f-7kc8c" Oct 07 14:01:33 crc kubenswrapper[4959]: I1007 14:01:33.957381 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" event={"ID":"4c502b70-d657-47d6-a703-4ecf3cc84d01","Type":"ContainerStarted","Data":"a8a2200da57e5e47c36da78d602f44fd8a60bcac7cad79714ee6b62fcf3956d8"} Oct 07 14:01:33 crc kubenswrapper[4959]: I1007 14:01:33.957485 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" Oct 07 14:01:33 crc kubenswrapper[4959]: I1007 14:01:33.991960 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" podStartSLOduration=3.533299777 podStartE2EDuration="3.991937829s" podCreationTimestamp="2025-10-07 14:01:30 +0000 UTC" firstStartedPulling="2025-10-07 14:01:31.707094268 +0000 UTC m=+973.790498593" lastFinishedPulling="2025-10-07 14:01:32.16573232 +0000 UTC m=+974.249136645" observedRunningTime="2025-10-07 14:01:33.990827221 +0000 UTC m=+976.074231546" watchObservedRunningTime="2025-10-07 14:01:33.991937829 +0000 UTC m=+976.075342154" Oct 07 14:01:34 crc kubenswrapper[4959]: I1007 14:01:34.009184 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-54c9499b4f-7kc8c" podStartSLOduration=3.390249882 podStartE2EDuration="4.00915972s" podCreationTimestamp="2025-10-07 14:01:30 +0000 UTC" firstStartedPulling="2025-10-07 14:01:31.598037885 +0000 UTC m=+973.681442220" lastFinishedPulling="2025-10-07 14:01:32.216947733 +0000 UTC m=+974.300352058" observedRunningTime="2025-10-07 14:01:33.975114457 +0000 UTC m=+976.058518782" watchObservedRunningTime="2025-10-07 14:01:34.00915972 +0000 UTC m=+976.092564055" Oct 07 14:01:34 crc kubenswrapper[4959]: I1007 14:01:34.292241 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:34 crc kubenswrapper[4959]: I1007 14:01:34.328612 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:34 crc kubenswrapper[4959]: I1007 14:01:34.965864 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.029987 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.219571 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.221305 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.225972 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.226171 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.226175 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.226251 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-hz99x" Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.267322 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.361189 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c76f0f7e-182d-40e0-b4bd-34076ba831e3-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c76f0f7e-182d-40e0-b4bd-34076ba831e3\") " pod="openstack/ovn-northd-0" Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.361233 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c76f0f7e-182d-40e0-b4bd-34076ba831e3-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c76f0f7e-182d-40e0-b4bd-34076ba831e3\") " pod="openstack/ovn-northd-0" Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.361541 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c76f0f7e-182d-40e0-b4bd-34076ba831e3-scripts\") pod \"ovn-northd-0\" (UID: \"c76f0f7e-182d-40e0-b4bd-34076ba831e3\") " pod="openstack/ovn-northd-0" Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.361719 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c76f0f7e-182d-40e0-b4bd-34076ba831e3-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c76f0f7e-182d-40e0-b4bd-34076ba831e3\") " pod="openstack/ovn-northd-0" Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.361884 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qw2v\" (UniqueName: \"kubernetes.io/projected/c76f0f7e-182d-40e0-b4bd-34076ba831e3-kube-api-access-8qw2v\") pod \"ovn-northd-0\" (UID: \"c76f0f7e-182d-40e0-b4bd-34076ba831e3\") " pod="openstack/ovn-northd-0" Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.361964 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c76f0f7e-182d-40e0-b4bd-34076ba831e3-config\") pod \"ovn-northd-0\" (UID: \"c76f0f7e-182d-40e0-b4bd-34076ba831e3\") " pod="openstack/ovn-northd-0" Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.362161 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c76f0f7e-182d-40e0-b4bd-34076ba831e3-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c76f0f7e-182d-40e0-b4bd-34076ba831e3\") " pod="openstack/ovn-northd-0" Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.463112 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c76f0f7e-182d-40e0-b4bd-34076ba831e3-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c76f0f7e-182d-40e0-b4bd-34076ba831e3\") " pod="openstack/ovn-northd-0" Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.463177 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c76f0f7e-182d-40e0-b4bd-34076ba831e3-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c76f0f7e-182d-40e0-b4bd-34076ba831e3\") " pod="openstack/ovn-northd-0" Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.463200 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c76f0f7e-182d-40e0-b4bd-34076ba831e3-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c76f0f7e-182d-40e0-b4bd-34076ba831e3\") " pod="openstack/ovn-northd-0" Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.463258 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c76f0f7e-182d-40e0-b4bd-34076ba831e3-scripts\") pod \"ovn-northd-0\" (UID: \"c76f0f7e-182d-40e0-b4bd-34076ba831e3\") " pod="openstack/ovn-northd-0" Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.463320 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c76f0f7e-182d-40e0-b4bd-34076ba831e3-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c76f0f7e-182d-40e0-b4bd-34076ba831e3\") " pod="openstack/ovn-northd-0" Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.463385 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qw2v\" (UniqueName: \"kubernetes.io/projected/c76f0f7e-182d-40e0-b4bd-34076ba831e3-kube-api-access-8qw2v\") pod \"ovn-northd-0\" (UID: \"c76f0f7e-182d-40e0-b4bd-34076ba831e3\") " pod="openstack/ovn-northd-0" Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.463433 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c76f0f7e-182d-40e0-b4bd-34076ba831e3-config\") pod \"ovn-northd-0\" (UID: \"c76f0f7e-182d-40e0-b4bd-34076ba831e3\") " pod="openstack/ovn-northd-0" Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.464400 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c76f0f7e-182d-40e0-b4bd-34076ba831e3-config\") pod \"ovn-northd-0\" (UID: \"c76f0f7e-182d-40e0-b4bd-34076ba831e3\") " pod="openstack/ovn-northd-0" Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.464533 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c76f0f7e-182d-40e0-b4bd-34076ba831e3-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c76f0f7e-182d-40e0-b4bd-34076ba831e3\") " pod="openstack/ovn-northd-0" Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.464711 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c76f0f7e-182d-40e0-b4bd-34076ba831e3-scripts\") pod \"ovn-northd-0\" (UID: \"c76f0f7e-182d-40e0-b4bd-34076ba831e3\") " pod="openstack/ovn-northd-0" Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.469870 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c76f0f7e-182d-40e0-b4bd-34076ba831e3-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c76f0f7e-182d-40e0-b4bd-34076ba831e3\") " pod="openstack/ovn-northd-0" Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.469998 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c76f0f7e-182d-40e0-b4bd-34076ba831e3-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c76f0f7e-182d-40e0-b4bd-34076ba831e3\") " pod="openstack/ovn-northd-0" Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.470244 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c76f0f7e-182d-40e0-b4bd-34076ba831e3-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c76f0f7e-182d-40e0-b4bd-34076ba831e3\") " pod="openstack/ovn-northd-0" Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.481672 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qw2v\" (UniqueName: \"kubernetes.io/projected/c76f0f7e-182d-40e0-b4bd-34076ba831e3-kube-api-access-8qw2v\") pod \"ovn-northd-0\" (UID: \"c76f0f7e-182d-40e0-b4bd-34076ba831e3\") " pod="openstack/ovn-northd-0" Oct 07 14:01:35 crc kubenswrapper[4959]: I1007 14:01:35.551031 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 07 14:01:36 crc kubenswrapper[4959]: I1007 14:01:35.999871 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 07 14:01:36 crc kubenswrapper[4959]: W1007 14:01:36.000905 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc76f0f7e_182d_40e0_b4bd_34076ba831e3.slice/crio-75fb40fb78b8ede27b2569dc0f7119a975b3d35b8a56887f0b0a1c7f8f16932b WatchSource:0}: Error finding container 75fb40fb78b8ede27b2569dc0f7119a975b3d35b8a56887f0b0a1c7f8f16932b: Status 404 returned error can't find the container with id 75fb40fb78b8ede27b2569dc0f7119a975b3d35b8a56887f0b0a1c7f8f16932b Oct 07 14:01:36 crc kubenswrapper[4959]: I1007 14:01:36.979276 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c76f0f7e-182d-40e0-b4bd-34076ba831e3","Type":"ContainerStarted","Data":"75fb40fb78b8ede27b2569dc0f7119a975b3d35b8a56887f0b0a1c7f8f16932b"} Oct 07 14:01:40 crc kubenswrapper[4959]: I1007 14:01:40.808279 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-54c9499b4f-7kc8c" Oct 07 14:01:40 crc kubenswrapper[4959]: I1007 14:01:40.860530 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" Oct 07 14:01:40 crc kubenswrapper[4959]: I1007 14:01:40.923052 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 07 14:01:40 crc kubenswrapper[4959]: I1007 14:01:40.923090 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 07 14:01:40 crc kubenswrapper[4959]: I1007 14:01:40.964498 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54c9499b4f-7kc8c"] Oct 07 14:01:41 crc kubenswrapper[4959]: I1007 14:01:41.017465 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c76f0f7e-182d-40e0-b4bd-34076ba831e3","Type":"ContainerStarted","Data":"1330a80ccd6d097e15486390157f2529320646211bfbc01a5a60be65e1e80828"} Oct 07 14:01:41 crc kubenswrapper[4959]: I1007 14:01:41.017534 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c76f0f7e-182d-40e0-b4bd-34076ba831e3","Type":"ContainerStarted","Data":"c1f7a844bdb13c881f7fc73ddd42c9051deccf38b863513bcf11027c7316a728"} Oct 07 14:01:41 crc kubenswrapper[4959]: I1007 14:01:41.017476 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-54c9499b4f-7kc8c" podUID="c729d934-aff5-4b38-a0de-aa8526fe1b05" containerName="dnsmasq-dns" containerID="cri-o://e448092192070ea048ed33d088ebd517a3f52143dcdff42e672bd7a34fcfe546" gracePeriod=10 Oct 07 14:01:41 crc kubenswrapper[4959]: I1007 14:01:41.018749 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 07 14:01:41 crc kubenswrapper[4959]: I1007 14:01:41.052212 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.310747417 podStartE2EDuration="6.052187765s" podCreationTimestamp="2025-10-07 14:01:35 +0000 UTC" firstStartedPulling="2025-10-07 14:01:36.002761343 +0000 UTC m=+978.086165668" lastFinishedPulling="2025-10-07 14:01:39.744201691 +0000 UTC m=+981.827606016" observedRunningTime="2025-10-07 14:01:41.047033495 +0000 UTC m=+983.130437810" watchObservedRunningTime="2025-10-07 14:01:41.052187765 +0000 UTC m=+983.135592090" Oct 07 14:01:41 crc kubenswrapper[4959]: I1007 14:01:41.143994 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 07 14:01:41 crc kubenswrapper[4959]: I1007 14:01:41.191730 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 07 14:01:41 crc kubenswrapper[4959]: I1007 14:01:41.301627 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:41 crc kubenswrapper[4959]: I1007 14:01:41.302456 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:41 crc kubenswrapper[4959]: I1007 14:01:41.343596 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:41 crc kubenswrapper[4959]: I1007 14:01:41.590566 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-jgf7k"] Oct 07 14:01:41 crc kubenswrapper[4959]: I1007 14:01:41.591792 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jgf7k" Oct 07 14:01:41 crc kubenswrapper[4959]: I1007 14:01:41.605061 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-jgf7k"] Oct 07 14:01:41 crc kubenswrapper[4959]: I1007 14:01:41.704811 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzq7k\" (UniqueName: \"kubernetes.io/projected/67166a6a-e8f0-479e-acda-1fec5161eac4-kube-api-access-xzq7k\") pod \"keystone-db-create-jgf7k\" (UID: \"67166a6a-e8f0-479e-acda-1fec5161eac4\") " pod="openstack/keystone-db-create-jgf7k" Oct 07 14:01:41 crc kubenswrapper[4959]: I1007 14:01:41.793272 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-sxjr4"] Oct 07 14:01:41 crc kubenswrapper[4959]: I1007 14:01:41.794716 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-sxjr4" Oct 07 14:01:41 crc kubenswrapper[4959]: I1007 14:01:41.801177 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-sxjr4"] Oct 07 14:01:41 crc kubenswrapper[4959]: I1007 14:01:41.806770 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzq7k\" (UniqueName: \"kubernetes.io/projected/67166a6a-e8f0-479e-acda-1fec5161eac4-kube-api-access-xzq7k\") pod \"keystone-db-create-jgf7k\" (UID: \"67166a6a-e8f0-479e-acda-1fec5161eac4\") " pod="openstack/keystone-db-create-jgf7k" Oct 07 14:01:41 crc kubenswrapper[4959]: I1007 14:01:41.858312 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzq7k\" (UniqueName: \"kubernetes.io/projected/67166a6a-e8f0-479e-acda-1fec5161eac4-kube-api-access-xzq7k\") pod \"keystone-db-create-jgf7k\" (UID: \"67166a6a-e8f0-479e-acda-1fec5161eac4\") " pod="openstack/keystone-db-create-jgf7k" Oct 07 14:01:41 crc kubenswrapper[4959]: I1007 14:01:41.908458 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkwq2\" (UniqueName: \"kubernetes.io/projected/bbe13445-5cad-4de1-aed4-6c95c41c83b2-kube-api-access-fkwq2\") pod \"placement-db-create-sxjr4\" (UID: \"bbe13445-5cad-4de1-aed4-6c95c41c83b2\") " pod="openstack/placement-db-create-sxjr4" Oct 07 14:01:41 crc kubenswrapper[4959]: I1007 14:01:41.912167 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jgf7k" Oct 07 14:01:42 crc kubenswrapper[4959]: I1007 14:01:42.010295 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkwq2\" (UniqueName: \"kubernetes.io/projected/bbe13445-5cad-4de1-aed4-6c95c41c83b2-kube-api-access-fkwq2\") pod \"placement-db-create-sxjr4\" (UID: \"bbe13445-5cad-4de1-aed4-6c95c41c83b2\") " pod="openstack/placement-db-create-sxjr4" Oct 07 14:01:42 crc kubenswrapper[4959]: I1007 14:01:42.037268 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkwq2\" (UniqueName: \"kubernetes.io/projected/bbe13445-5cad-4de1-aed4-6c95c41c83b2-kube-api-access-fkwq2\") pod \"placement-db-create-sxjr4\" (UID: \"bbe13445-5cad-4de1-aed4-6c95c41c83b2\") " pod="openstack/placement-db-create-sxjr4" Oct 07 14:01:42 crc kubenswrapper[4959]: I1007 14:01:42.043981 4959 generic.go:334] "Generic (PLEG): container finished" podID="c729d934-aff5-4b38-a0de-aa8526fe1b05" containerID="e448092192070ea048ed33d088ebd517a3f52143dcdff42e672bd7a34fcfe546" exitCode=0 Oct 07 14:01:42 crc kubenswrapper[4959]: I1007 14:01:42.045307 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c9499b4f-7kc8c" event={"ID":"c729d934-aff5-4b38-a0de-aa8526fe1b05","Type":"ContainerDied","Data":"e448092192070ea048ed33d088ebd517a3f52143dcdff42e672bd7a34fcfe546"} Oct 07 14:01:42 crc kubenswrapper[4959]: I1007 14:01:42.119004 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-4k9wm"] Oct 07 14:01:42 crc kubenswrapper[4959]: I1007 14:01:42.121109 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-4k9wm" Oct 07 14:01:42 crc kubenswrapper[4959]: I1007 14:01:42.125004 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-4k9wm"] Oct 07 14:01:42 crc kubenswrapper[4959]: I1007 14:01:42.135383 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 07 14:01:42 crc kubenswrapper[4959]: I1007 14:01:42.156190 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-sxjr4" Oct 07 14:01:42 crc kubenswrapper[4959]: I1007 14:01:42.222160 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cclfc\" (UniqueName: \"kubernetes.io/projected/b2ed38ec-6665-4e4f-9a10-1eb8c958a306-kube-api-access-cclfc\") pod \"glance-db-create-4k9wm\" (UID: \"b2ed38ec-6665-4e4f-9a10-1eb8c958a306\") " pod="openstack/glance-db-create-4k9wm" Oct 07 14:01:42 crc kubenswrapper[4959]: I1007 14:01:42.323480 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cclfc\" (UniqueName: \"kubernetes.io/projected/b2ed38ec-6665-4e4f-9a10-1eb8c958a306-kube-api-access-cclfc\") pod \"glance-db-create-4k9wm\" (UID: \"b2ed38ec-6665-4e4f-9a10-1eb8c958a306\") " pod="openstack/glance-db-create-4k9wm" Oct 07 14:01:42 crc kubenswrapper[4959]: I1007 14:01:42.342969 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cclfc\" (UniqueName: \"kubernetes.io/projected/b2ed38ec-6665-4e4f-9a10-1eb8c958a306-kube-api-access-cclfc\") pod \"glance-db-create-4k9wm\" (UID: \"b2ed38ec-6665-4e4f-9a10-1eb8c958a306\") " pod="openstack/glance-db-create-4k9wm" Oct 07 14:01:42 crc kubenswrapper[4959]: I1007 14:01:42.379494 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-jgf7k"] Oct 07 14:01:42 crc kubenswrapper[4959]: W1007 14:01:42.396906 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67166a6a_e8f0_479e_acda_1fec5161eac4.slice/crio-5ffcc48ebfa5537082148d581855ec310c5bfde888de1096ef7c924ae8b304df WatchSource:0}: Error finding container 5ffcc48ebfa5537082148d581855ec310c5bfde888de1096ef7c924ae8b304df: Status 404 returned error can't find the container with id 5ffcc48ebfa5537082148d581855ec310c5bfde888de1096ef7c924ae8b304df Oct 07 14:01:42 crc kubenswrapper[4959]: I1007 14:01:42.445884 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-4k9wm" Oct 07 14:01:42 crc kubenswrapper[4959]: I1007 14:01:42.623599 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-sxjr4"] Oct 07 14:01:42 crc kubenswrapper[4959]: I1007 14:01:42.637481 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54c9499b4f-7kc8c" Oct 07 14:01:42 crc kubenswrapper[4959]: W1007 14:01:42.644206 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbbe13445_5cad_4de1_aed4_6c95c41c83b2.slice/crio-bf5f98b9401ccaf266bfb41cbe8790a6c9bf59408248a96903b6a7c2bbc2d0d5 WatchSource:0}: Error finding container bf5f98b9401ccaf266bfb41cbe8790a6c9bf59408248a96903b6a7c2bbc2d0d5: Status 404 returned error can't find the container with id bf5f98b9401ccaf266bfb41cbe8790a6c9bf59408248a96903b6a7c2bbc2d0d5 Oct 07 14:01:42 crc kubenswrapper[4959]: I1007 14:01:42.732035 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c729d934-aff5-4b38-a0de-aa8526fe1b05-config\") pod \"c729d934-aff5-4b38-a0de-aa8526fe1b05\" (UID: \"c729d934-aff5-4b38-a0de-aa8526fe1b05\") " Oct 07 14:01:42 crc kubenswrapper[4959]: I1007 14:01:42.732541 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wd85k\" (UniqueName: \"kubernetes.io/projected/c729d934-aff5-4b38-a0de-aa8526fe1b05-kube-api-access-wd85k\") pod \"c729d934-aff5-4b38-a0de-aa8526fe1b05\" (UID: \"c729d934-aff5-4b38-a0de-aa8526fe1b05\") " Oct 07 14:01:42 crc kubenswrapper[4959]: I1007 14:01:42.732618 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c729d934-aff5-4b38-a0de-aa8526fe1b05-dns-svc\") pod \"c729d934-aff5-4b38-a0de-aa8526fe1b05\" (UID: \"c729d934-aff5-4b38-a0de-aa8526fe1b05\") " Oct 07 14:01:42 crc kubenswrapper[4959]: I1007 14:01:42.732766 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c729d934-aff5-4b38-a0de-aa8526fe1b05-ovsdbserver-sb\") pod \"c729d934-aff5-4b38-a0de-aa8526fe1b05\" (UID: \"c729d934-aff5-4b38-a0de-aa8526fe1b05\") " Oct 07 14:01:42 crc kubenswrapper[4959]: I1007 14:01:42.747182 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c729d934-aff5-4b38-a0de-aa8526fe1b05-kube-api-access-wd85k" (OuterVolumeSpecName: "kube-api-access-wd85k") pod "c729d934-aff5-4b38-a0de-aa8526fe1b05" (UID: "c729d934-aff5-4b38-a0de-aa8526fe1b05"). InnerVolumeSpecName "kube-api-access-wd85k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:01:42 crc kubenswrapper[4959]: I1007 14:01:42.774435 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c729d934-aff5-4b38-a0de-aa8526fe1b05-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c729d934-aff5-4b38-a0de-aa8526fe1b05" (UID: "c729d934-aff5-4b38-a0de-aa8526fe1b05"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:01:42 crc kubenswrapper[4959]: I1007 14:01:42.796395 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c729d934-aff5-4b38-a0de-aa8526fe1b05-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c729d934-aff5-4b38-a0de-aa8526fe1b05" (UID: "c729d934-aff5-4b38-a0de-aa8526fe1b05"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:01:42 crc kubenswrapper[4959]: I1007 14:01:42.803442 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c729d934-aff5-4b38-a0de-aa8526fe1b05-config" (OuterVolumeSpecName: "config") pod "c729d934-aff5-4b38-a0de-aa8526fe1b05" (UID: "c729d934-aff5-4b38-a0de-aa8526fe1b05"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:01:42 crc kubenswrapper[4959]: I1007 14:01:42.836776 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c729d934-aff5-4b38-a0de-aa8526fe1b05-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 14:01:42 crc kubenswrapper[4959]: I1007 14:01:42.836833 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c729d934-aff5-4b38-a0de-aa8526fe1b05-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 14:01:42 crc kubenswrapper[4959]: I1007 14:01:42.836845 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c729d934-aff5-4b38-a0de-aa8526fe1b05-config\") on node \"crc\" DevicePath \"\"" Oct 07 14:01:42 crc kubenswrapper[4959]: I1007 14:01:42.836854 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wd85k\" (UniqueName: \"kubernetes.io/projected/c729d934-aff5-4b38-a0de-aa8526fe1b05-kube-api-access-wd85k\") on node \"crc\" DevicePath \"\"" Oct 07 14:01:42 crc kubenswrapper[4959]: I1007 14:01:42.983265 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-4k9wm"] Oct 07 14:01:43 crc kubenswrapper[4959]: I1007 14:01:43.053161 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-4k9wm" event={"ID":"b2ed38ec-6665-4e4f-9a10-1eb8c958a306","Type":"ContainerStarted","Data":"72e890905e175a4ce1be3babdf4afd898038c3369a695d880275eb5ff1268c4f"} Oct 07 14:01:43 crc kubenswrapper[4959]: I1007 14:01:43.056379 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jgf7k" event={"ID":"67166a6a-e8f0-479e-acda-1fec5161eac4","Type":"ContainerStarted","Data":"5ffcc48ebfa5537082148d581855ec310c5bfde888de1096ef7c924ae8b304df"} Oct 07 14:01:43 crc kubenswrapper[4959]: I1007 14:01:43.058018 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-sxjr4" event={"ID":"bbe13445-5cad-4de1-aed4-6c95c41c83b2","Type":"ContainerStarted","Data":"bf5f98b9401ccaf266bfb41cbe8790a6c9bf59408248a96903b6a7c2bbc2d0d5"} Oct 07 14:01:43 crc kubenswrapper[4959]: I1007 14:01:43.060354 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c9499b4f-7kc8c" event={"ID":"c729d934-aff5-4b38-a0de-aa8526fe1b05","Type":"ContainerDied","Data":"c26a1e07085b7c468a462dfdbf509d16df7ee5a20e17e49dbb39c7669f27d766"} Oct 07 14:01:43 crc kubenswrapper[4959]: I1007 14:01:43.060393 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54c9499b4f-7kc8c" Oct 07 14:01:43 crc kubenswrapper[4959]: I1007 14:01:43.060420 4959 scope.go:117] "RemoveContainer" containerID="e448092192070ea048ed33d088ebd517a3f52143dcdff42e672bd7a34fcfe546" Oct 07 14:01:43 crc kubenswrapper[4959]: I1007 14:01:43.076623 4959 scope.go:117] "RemoveContainer" containerID="ce7d0f7bf1c0c709d5f3cc8e1e688d6ea6647bd5dd8c76d503c55b699afb79c1" Oct 07 14:01:43 crc kubenswrapper[4959]: I1007 14:01:43.124160 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54c9499b4f-7kc8c"] Oct 07 14:01:43 crc kubenswrapper[4959]: I1007 14:01:43.130595 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54c9499b4f-7kc8c"] Oct 07 14:01:44 crc kubenswrapper[4959]: I1007 14:01:44.069682 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jgf7k" event={"ID":"67166a6a-e8f0-479e-acda-1fec5161eac4","Type":"ContainerStarted","Data":"ee3fd2485575e455c60f1fb347f4e962b59a424640d5ff51f426bfe326a4546b"} Oct 07 14:01:44 crc kubenswrapper[4959]: I1007 14:01:44.071610 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-sxjr4" event={"ID":"bbe13445-5cad-4de1-aed4-6c95c41c83b2","Type":"ContainerStarted","Data":"852d37b59bb0d41d1e34f8f3b61f053265e9f6cd4f9ae4a11c4f2fd232721040"} Oct 07 14:01:44 crc kubenswrapper[4959]: I1007 14:01:44.074325 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-4k9wm" event={"ID":"b2ed38ec-6665-4e4f-9a10-1eb8c958a306","Type":"ContainerStarted","Data":"f652c5f885e6d1fdb8237b16d9ad7f2ecfd9fd09d3af2513f73f898bf90cadff"} Oct 07 14:01:44 crc kubenswrapper[4959]: I1007 14:01:44.663354 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c729d934-aff5-4b38-a0de-aa8526fe1b05" path="/var/lib/kubelet/pods/c729d934-aff5-4b38-a0de-aa8526fe1b05/volumes" Oct 07 14:01:45 crc kubenswrapper[4959]: I1007 14:01:45.087149 4959 generic.go:334] "Generic (PLEG): container finished" podID="67166a6a-e8f0-479e-acda-1fec5161eac4" containerID="ee3fd2485575e455c60f1fb347f4e962b59a424640d5ff51f426bfe326a4546b" exitCode=0 Oct 07 14:01:45 crc kubenswrapper[4959]: I1007 14:01:45.087255 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jgf7k" event={"ID":"67166a6a-e8f0-479e-acda-1fec5161eac4","Type":"ContainerDied","Data":"ee3fd2485575e455c60f1fb347f4e962b59a424640d5ff51f426bfe326a4546b"} Oct 07 14:01:45 crc kubenswrapper[4959]: I1007 14:01:45.101300 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-sxjr4" podStartSLOduration=4.101279591 podStartE2EDuration="4.101279591s" podCreationTimestamp="2025-10-07 14:01:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:01:45.09883753 +0000 UTC m=+987.182241875" watchObservedRunningTime="2025-10-07 14:01:45.101279591 +0000 UTC m=+987.184683916" Oct 07 14:01:45 crc kubenswrapper[4959]: I1007 14:01:45.117371 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-4k9wm" podStartSLOduration=3.117350534 podStartE2EDuration="3.117350534s" podCreationTimestamp="2025-10-07 14:01:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:01:45.112930273 +0000 UTC m=+987.196334608" watchObservedRunningTime="2025-10-07 14:01:45.117350534 +0000 UTC m=+987.200754859" Oct 07 14:01:46 crc kubenswrapper[4959]: I1007 14:01:46.098136 4959 generic.go:334] "Generic (PLEG): container finished" podID="b2ed38ec-6665-4e4f-9a10-1eb8c958a306" containerID="f652c5f885e6d1fdb8237b16d9ad7f2ecfd9fd09d3af2513f73f898bf90cadff" exitCode=0 Oct 07 14:01:46 crc kubenswrapper[4959]: I1007 14:01:46.098254 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-4k9wm" event={"ID":"b2ed38ec-6665-4e4f-9a10-1eb8c958a306","Type":"ContainerDied","Data":"f652c5f885e6d1fdb8237b16d9ad7f2ecfd9fd09d3af2513f73f898bf90cadff"} Oct 07 14:01:46 crc kubenswrapper[4959]: I1007 14:01:46.100458 4959 generic.go:334] "Generic (PLEG): container finished" podID="bbe13445-5cad-4de1-aed4-6c95c41c83b2" containerID="852d37b59bb0d41d1e34f8f3b61f053265e9f6cd4f9ae4a11c4f2fd232721040" exitCode=0 Oct 07 14:01:46 crc kubenswrapper[4959]: I1007 14:01:46.100492 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-sxjr4" event={"ID":"bbe13445-5cad-4de1-aed4-6c95c41c83b2","Type":"ContainerDied","Data":"852d37b59bb0d41d1e34f8f3b61f053265e9f6cd4f9ae4a11c4f2fd232721040"} Oct 07 14:01:46 crc kubenswrapper[4959]: I1007 14:01:46.448369 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jgf7k" Oct 07 14:01:46 crc kubenswrapper[4959]: I1007 14:01:46.598152 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzq7k\" (UniqueName: \"kubernetes.io/projected/67166a6a-e8f0-479e-acda-1fec5161eac4-kube-api-access-xzq7k\") pod \"67166a6a-e8f0-479e-acda-1fec5161eac4\" (UID: \"67166a6a-e8f0-479e-acda-1fec5161eac4\") " Oct 07 14:01:46 crc kubenswrapper[4959]: I1007 14:01:46.606510 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67166a6a-e8f0-479e-acda-1fec5161eac4-kube-api-access-xzq7k" (OuterVolumeSpecName: "kube-api-access-xzq7k") pod "67166a6a-e8f0-479e-acda-1fec5161eac4" (UID: "67166a6a-e8f0-479e-acda-1fec5161eac4"). InnerVolumeSpecName "kube-api-access-xzq7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:01:46 crc kubenswrapper[4959]: I1007 14:01:46.699835 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzq7k\" (UniqueName: \"kubernetes.io/projected/67166a6a-e8f0-479e-acda-1fec5161eac4-kube-api-access-xzq7k\") on node \"crc\" DevicePath \"\"" Oct 07 14:01:47 crc kubenswrapper[4959]: I1007 14:01:47.109024 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jgf7k" event={"ID":"67166a6a-e8f0-479e-acda-1fec5161eac4","Type":"ContainerDied","Data":"5ffcc48ebfa5537082148d581855ec310c5bfde888de1096ef7c924ae8b304df"} Oct 07 14:01:47 crc kubenswrapper[4959]: I1007 14:01:47.109441 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ffcc48ebfa5537082148d581855ec310c5bfde888de1096ef7c924ae8b304df" Oct 07 14:01:47 crc kubenswrapper[4959]: I1007 14:01:47.109091 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jgf7k" Oct 07 14:01:47 crc kubenswrapper[4959]: I1007 14:01:47.545924 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-4k9wm" Oct 07 14:01:47 crc kubenswrapper[4959]: I1007 14:01:47.558414 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-sxjr4" Oct 07 14:01:47 crc kubenswrapper[4959]: I1007 14:01:47.716998 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cclfc\" (UniqueName: \"kubernetes.io/projected/b2ed38ec-6665-4e4f-9a10-1eb8c958a306-kube-api-access-cclfc\") pod \"b2ed38ec-6665-4e4f-9a10-1eb8c958a306\" (UID: \"b2ed38ec-6665-4e4f-9a10-1eb8c958a306\") " Oct 07 14:01:47 crc kubenswrapper[4959]: I1007 14:01:47.717085 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkwq2\" (UniqueName: \"kubernetes.io/projected/bbe13445-5cad-4de1-aed4-6c95c41c83b2-kube-api-access-fkwq2\") pod \"bbe13445-5cad-4de1-aed4-6c95c41c83b2\" (UID: \"bbe13445-5cad-4de1-aed4-6c95c41c83b2\") " Oct 07 14:01:47 crc kubenswrapper[4959]: I1007 14:01:47.721556 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbe13445-5cad-4de1-aed4-6c95c41c83b2-kube-api-access-fkwq2" (OuterVolumeSpecName: "kube-api-access-fkwq2") pod "bbe13445-5cad-4de1-aed4-6c95c41c83b2" (UID: "bbe13445-5cad-4de1-aed4-6c95c41c83b2"). InnerVolumeSpecName "kube-api-access-fkwq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:01:47 crc kubenswrapper[4959]: I1007 14:01:47.723995 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2ed38ec-6665-4e4f-9a10-1eb8c958a306-kube-api-access-cclfc" (OuterVolumeSpecName: "kube-api-access-cclfc") pod "b2ed38ec-6665-4e4f-9a10-1eb8c958a306" (UID: "b2ed38ec-6665-4e4f-9a10-1eb8c958a306"). InnerVolumeSpecName "kube-api-access-cclfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:01:47 crc kubenswrapper[4959]: I1007 14:01:47.819222 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cclfc\" (UniqueName: \"kubernetes.io/projected/b2ed38ec-6665-4e4f-9a10-1eb8c958a306-kube-api-access-cclfc\") on node \"crc\" DevicePath \"\"" Oct 07 14:01:47 crc kubenswrapper[4959]: I1007 14:01:47.819256 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkwq2\" (UniqueName: \"kubernetes.io/projected/bbe13445-5cad-4de1-aed4-6c95c41c83b2-kube-api-access-fkwq2\") on node \"crc\" DevicePath \"\"" Oct 07 14:01:48 crc kubenswrapper[4959]: I1007 14:01:48.118398 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-sxjr4" event={"ID":"bbe13445-5cad-4de1-aed4-6c95c41c83b2","Type":"ContainerDied","Data":"bf5f98b9401ccaf266bfb41cbe8790a6c9bf59408248a96903b6a7c2bbc2d0d5"} Oct 07 14:01:48 crc kubenswrapper[4959]: I1007 14:01:48.118420 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-sxjr4" Oct 07 14:01:48 crc kubenswrapper[4959]: I1007 14:01:48.118446 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf5f98b9401ccaf266bfb41cbe8790a6c9bf59408248a96903b6a7c2bbc2d0d5" Oct 07 14:01:48 crc kubenswrapper[4959]: I1007 14:01:48.120811 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-4k9wm" event={"ID":"b2ed38ec-6665-4e4f-9a10-1eb8c958a306","Type":"ContainerDied","Data":"72e890905e175a4ce1be3babdf4afd898038c3369a695d880275eb5ff1268c4f"} Oct 07 14:01:48 crc kubenswrapper[4959]: I1007 14:01:48.120832 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72e890905e175a4ce1be3babdf4afd898038c3369a695d880275eb5ff1268c4f" Oct 07 14:01:48 crc kubenswrapper[4959]: I1007 14:01:48.120899 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-4k9wm" Oct 07 14:01:50 crc kubenswrapper[4959]: I1007 14:01:50.643847 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 07 14:01:52 crc kubenswrapper[4959]: I1007 14:01:52.222454 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-6996-account-create-pvzth"] Oct 07 14:01:52 crc kubenswrapper[4959]: E1007 14:01:52.222934 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2ed38ec-6665-4e4f-9a10-1eb8c958a306" containerName="mariadb-database-create" Oct 07 14:01:52 crc kubenswrapper[4959]: I1007 14:01:52.222946 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2ed38ec-6665-4e4f-9a10-1eb8c958a306" containerName="mariadb-database-create" Oct 07 14:01:52 crc kubenswrapper[4959]: E1007 14:01:52.222961 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbe13445-5cad-4de1-aed4-6c95c41c83b2" containerName="mariadb-database-create" Oct 07 14:01:52 crc kubenswrapper[4959]: I1007 14:01:52.222967 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbe13445-5cad-4de1-aed4-6c95c41c83b2" containerName="mariadb-database-create" Oct 07 14:01:52 crc kubenswrapper[4959]: E1007 14:01:52.222980 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67166a6a-e8f0-479e-acda-1fec5161eac4" containerName="mariadb-database-create" Oct 07 14:01:52 crc kubenswrapper[4959]: I1007 14:01:52.222986 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="67166a6a-e8f0-479e-acda-1fec5161eac4" containerName="mariadb-database-create" Oct 07 14:01:52 crc kubenswrapper[4959]: E1007 14:01:52.223005 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c729d934-aff5-4b38-a0de-aa8526fe1b05" containerName="init" Oct 07 14:01:52 crc kubenswrapper[4959]: I1007 14:01:52.223010 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c729d934-aff5-4b38-a0de-aa8526fe1b05" containerName="init" Oct 07 14:01:52 crc kubenswrapper[4959]: E1007 14:01:52.223022 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c729d934-aff5-4b38-a0de-aa8526fe1b05" containerName="dnsmasq-dns" Oct 07 14:01:52 crc kubenswrapper[4959]: I1007 14:01:52.223027 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c729d934-aff5-4b38-a0de-aa8526fe1b05" containerName="dnsmasq-dns" Oct 07 14:01:52 crc kubenswrapper[4959]: I1007 14:01:52.223163 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbe13445-5cad-4de1-aed4-6c95c41c83b2" containerName="mariadb-database-create" Oct 07 14:01:52 crc kubenswrapper[4959]: I1007 14:01:52.223173 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2ed38ec-6665-4e4f-9a10-1eb8c958a306" containerName="mariadb-database-create" Oct 07 14:01:52 crc kubenswrapper[4959]: I1007 14:01:52.223185 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="67166a6a-e8f0-479e-acda-1fec5161eac4" containerName="mariadb-database-create" Oct 07 14:01:52 crc kubenswrapper[4959]: I1007 14:01:52.223193 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="c729d934-aff5-4b38-a0de-aa8526fe1b05" containerName="dnsmasq-dns" Oct 07 14:01:52 crc kubenswrapper[4959]: I1007 14:01:52.223669 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6996-account-create-pvzth" Oct 07 14:01:52 crc kubenswrapper[4959]: I1007 14:01:52.226186 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 07 14:01:52 crc kubenswrapper[4959]: I1007 14:01:52.230010 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-6996-account-create-pvzth"] Oct 07 14:01:52 crc kubenswrapper[4959]: I1007 14:01:52.294770 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twhc4\" (UniqueName: \"kubernetes.io/projected/e8161816-d8b5-4e8a-a132-3ae55dcff880-kube-api-access-twhc4\") pod \"glance-6996-account-create-pvzth\" (UID: \"e8161816-d8b5-4e8a-a132-3ae55dcff880\") " pod="openstack/glance-6996-account-create-pvzth" Oct 07 14:01:52 crc kubenswrapper[4959]: I1007 14:01:52.395871 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twhc4\" (UniqueName: \"kubernetes.io/projected/e8161816-d8b5-4e8a-a132-3ae55dcff880-kube-api-access-twhc4\") pod \"glance-6996-account-create-pvzth\" (UID: \"e8161816-d8b5-4e8a-a132-3ae55dcff880\") " pod="openstack/glance-6996-account-create-pvzth" Oct 07 14:01:52 crc kubenswrapper[4959]: I1007 14:01:52.415399 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twhc4\" (UniqueName: \"kubernetes.io/projected/e8161816-d8b5-4e8a-a132-3ae55dcff880-kube-api-access-twhc4\") pod \"glance-6996-account-create-pvzth\" (UID: \"e8161816-d8b5-4e8a-a132-3ae55dcff880\") " pod="openstack/glance-6996-account-create-pvzth" Oct 07 14:01:52 crc kubenswrapper[4959]: I1007 14:01:52.544294 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6996-account-create-pvzth" Oct 07 14:01:52 crc kubenswrapper[4959]: I1007 14:01:52.950930 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-6996-account-create-pvzth"] Oct 07 14:01:52 crc kubenswrapper[4959]: W1007 14:01:52.994081 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8161816_d8b5_4e8a_a132_3ae55dcff880.slice/crio-bcde11c12eed9ec5f5fa12228753eadc85ad123a2735f4a9651cb5cf2eda8c2d WatchSource:0}: Error finding container bcde11c12eed9ec5f5fa12228753eadc85ad123a2735f4a9651cb5cf2eda8c2d: Status 404 returned error can't find the container with id bcde11c12eed9ec5f5fa12228753eadc85ad123a2735f4a9651cb5cf2eda8c2d Oct 07 14:01:53 crc kubenswrapper[4959]: I1007 14:01:53.160611 4959 generic.go:334] "Generic (PLEG): container finished" podID="90af7ce4-adf2-4722-a726-cc4d7d29056f" containerID="b6140a1530482e80a1f682355efe01fbc1dd47f08e5fd863d3a4a6fcf5355ee8" exitCode=0 Oct 07 14:01:53 crc kubenswrapper[4959]: I1007 14:01:53.160691 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"90af7ce4-adf2-4722-a726-cc4d7d29056f","Type":"ContainerDied","Data":"b6140a1530482e80a1f682355efe01fbc1dd47f08e5fd863d3a4a6fcf5355ee8"} Oct 07 14:01:53 crc kubenswrapper[4959]: I1007 14:01:53.162622 4959 generic.go:334] "Generic (PLEG): container finished" podID="e206a4fa-c178-4f85-be59-46edf7e78833" containerID="a9137451d6a4f2897ca96d09e23e311123bffd7256a6e6f4b2714386a1d4c45c" exitCode=0 Oct 07 14:01:53 crc kubenswrapper[4959]: I1007 14:01:53.162689 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e206a4fa-c178-4f85-be59-46edf7e78833","Type":"ContainerDied","Data":"a9137451d6a4f2897ca96d09e23e311123bffd7256a6e6f4b2714386a1d4c45c"} Oct 07 14:01:53 crc kubenswrapper[4959]: I1007 14:01:53.165053 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6996-account-create-pvzth" event={"ID":"e8161816-d8b5-4e8a-a132-3ae55dcff880","Type":"ContainerStarted","Data":"bcde11c12eed9ec5f5fa12228753eadc85ad123a2735f4a9651cb5cf2eda8c2d"} Oct 07 14:01:55 crc kubenswrapper[4959]: I1007 14:01:55.184433 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e206a4fa-c178-4f85-be59-46edf7e78833","Type":"ContainerStarted","Data":"ba386e7a023b2ed0780a071418693e842c7589ef16170ddd067523a4be613e71"} Oct 07 14:01:55 crc kubenswrapper[4959]: I1007 14:01:55.185002 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:01:55 crc kubenswrapper[4959]: I1007 14:01:55.186804 4959 generic.go:334] "Generic (PLEG): container finished" podID="e8161816-d8b5-4e8a-a132-3ae55dcff880" containerID="16de1daa04529c2509767c249c3c96e41a4928eab63f324fd5b3fd33f8c4d01b" exitCode=0 Oct 07 14:01:55 crc kubenswrapper[4959]: I1007 14:01:55.186866 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6996-account-create-pvzth" event={"ID":"e8161816-d8b5-4e8a-a132-3ae55dcff880","Type":"ContainerDied","Data":"16de1daa04529c2509767c249c3c96e41a4928eab63f324fd5b3fd33f8c4d01b"} Oct 07 14:01:55 crc kubenswrapper[4959]: I1007 14:01:55.188919 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"90af7ce4-adf2-4722-a726-cc4d7d29056f","Type":"ContainerStarted","Data":"0a122a18fbfa2dec9e05dd83c778908878ede8e5d5c924fd00f613322872d7e8"} Oct 07 14:01:55 crc kubenswrapper[4959]: I1007 14:01:55.189095 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 07 14:01:55 crc kubenswrapper[4959]: I1007 14:01:55.214409 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=39.239409015 podStartE2EDuration="59.214390382s" podCreationTimestamp="2025-10-07 14:00:56 +0000 UTC" firstStartedPulling="2025-10-07 14:00:59.179480245 +0000 UTC m=+941.262884570" lastFinishedPulling="2025-10-07 14:01:19.154461622 +0000 UTC m=+961.237865937" observedRunningTime="2025-10-07 14:01:55.208592436 +0000 UTC m=+997.291996761" watchObservedRunningTime="2025-10-07 14:01:55.214390382 +0000 UTC m=+997.297794707" Oct 07 14:01:56 crc kubenswrapper[4959]: I1007 14:01:56.512418 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6996-account-create-pvzth" Oct 07 14:01:56 crc kubenswrapper[4959]: I1007 14:01:56.528041 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=39.55583939 podStartE2EDuration="59.528021837s" podCreationTimestamp="2025-10-07 14:00:57 +0000 UTC" firstStartedPulling="2025-10-07 14:00:59.217274472 +0000 UTC m=+941.300678797" lastFinishedPulling="2025-10-07 14:01:19.189456919 +0000 UTC m=+961.272861244" observedRunningTime="2025-10-07 14:01:55.24466548 +0000 UTC m=+997.328069805" watchObservedRunningTime="2025-10-07 14:01:56.528021837 +0000 UTC m=+998.611426162" Oct 07 14:01:56 crc kubenswrapper[4959]: I1007 14:01:56.657436 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twhc4\" (UniqueName: \"kubernetes.io/projected/e8161816-d8b5-4e8a-a132-3ae55dcff880-kube-api-access-twhc4\") pod \"e8161816-d8b5-4e8a-a132-3ae55dcff880\" (UID: \"e8161816-d8b5-4e8a-a132-3ae55dcff880\") " Oct 07 14:01:56 crc kubenswrapper[4959]: I1007 14:01:56.663381 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8161816-d8b5-4e8a-a132-3ae55dcff880-kube-api-access-twhc4" (OuterVolumeSpecName: "kube-api-access-twhc4") pod "e8161816-d8b5-4e8a-a132-3ae55dcff880" (UID: "e8161816-d8b5-4e8a-a132-3ae55dcff880"). InnerVolumeSpecName "kube-api-access-twhc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:01:56 crc kubenswrapper[4959]: I1007 14:01:56.759601 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twhc4\" (UniqueName: \"kubernetes.io/projected/e8161816-d8b5-4e8a-a132-3ae55dcff880-kube-api-access-twhc4\") on node \"crc\" DevicePath \"\"" Oct 07 14:01:57 crc kubenswrapper[4959]: I1007 14:01:57.205880 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6996-account-create-pvzth" event={"ID":"e8161816-d8b5-4e8a-a132-3ae55dcff880","Type":"ContainerDied","Data":"bcde11c12eed9ec5f5fa12228753eadc85ad123a2735f4a9651cb5cf2eda8c2d"} Oct 07 14:01:57 crc kubenswrapper[4959]: I1007 14:01:57.206183 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bcde11c12eed9ec5f5fa12228753eadc85ad123a2735f4a9651cb5cf2eda8c2d" Oct 07 14:01:57 crc kubenswrapper[4959]: I1007 14:01:57.205969 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6996-account-create-pvzth" Oct 07 14:01:57 crc kubenswrapper[4959]: I1007 14:01:57.286278 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-q5v9v" podUID="603a8fc2-2eac-47b7-8e97-cd117033a313" containerName="ovn-controller" probeResult="failure" output=< Oct 07 14:01:57 crc kubenswrapper[4959]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 07 14:01:57 crc kubenswrapper[4959]: > Oct 07 14:02:01 crc kubenswrapper[4959]: I1007 14:02:01.578983 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-c1a2-account-create-xbdrw"] Oct 07 14:02:01 crc kubenswrapper[4959]: E1007 14:02:01.579700 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8161816-d8b5-4e8a-a132-3ae55dcff880" containerName="mariadb-account-create" Oct 07 14:02:01 crc kubenswrapper[4959]: I1007 14:02:01.579718 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8161816-d8b5-4e8a-a132-3ae55dcff880" containerName="mariadb-account-create" Oct 07 14:02:01 crc kubenswrapper[4959]: I1007 14:02:01.579912 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8161816-d8b5-4e8a-a132-3ae55dcff880" containerName="mariadb-account-create" Oct 07 14:02:01 crc kubenswrapper[4959]: I1007 14:02:01.580707 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c1a2-account-create-xbdrw" Oct 07 14:02:01 crc kubenswrapper[4959]: I1007 14:02:01.583268 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 07 14:02:01 crc kubenswrapper[4959]: I1007 14:02:01.585834 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-c1a2-account-create-xbdrw"] Oct 07 14:02:01 crc kubenswrapper[4959]: I1007 14:02:01.737503 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7l95\" (UniqueName: \"kubernetes.io/projected/ccf90649-6706-47f5-a899-f8674a14efcd-kube-api-access-p7l95\") pod \"keystone-c1a2-account-create-xbdrw\" (UID: \"ccf90649-6706-47f5-a899-f8674a14efcd\") " pod="openstack/keystone-c1a2-account-create-xbdrw" Oct 07 14:02:01 crc kubenswrapper[4959]: I1007 14:02:01.839823 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7l95\" (UniqueName: \"kubernetes.io/projected/ccf90649-6706-47f5-a899-f8674a14efcd-kube-api-access-p7l95\") pod \"keystone-c1a2-account-create-xbdrw\" (UID: \"ccf90649-6706-47f5-a899-f8674a14efcd\") " pod="openstack/keystone-c1a2-account-create-xbdrw" Oct 07 14:02:01 crc kubenswrapper[4959]: I1007 14:02:01.865393 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7l95\" (UniqueName: \"kubernetes.io/projected/ccf90649-6706-47f5-a899-f8674a14efcd-kube-api-access-p7l95\") pod \"keystone-c1a2-account-create-xbdrw\" (UID: \"ccf90649-6706-47f5-a899-f8674a14efcd\") " pod="openstack/keystone-c1a2-account-create-xbdrw" Oct 07 14:02:01 crc kubenswrapper[4959]: I1007 14:02:01.900396 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c1a2-account-create-xbdrw" Oct 07 14:02:01 crc kubenswrapper[4959]: I1007 14:02:01.914535 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-43c9-account-create-m7dhb"] Oct 07 14:02:01 crc kubenswrapper[4959]: I1007 14:02:01.915612 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-43c9-account-create-m7dhb" Oct 07 14:02:01 crc kubenswrapper[4959]: W1007 14:02:01.921160 4959 reflector.go:561] object-"openstack"/"placement-db-secret": failed to list *v1.Secret: secrets "placement-db-secret" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 07 14:02:01 crc kubenswrapper[4959]: E1007 14:02:01.921384 4959 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"placement-db-secret\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"placement-db-secret\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 07 14:02:01 crc kubenswrapper[4959]: I1007 14:02:01.925324 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-43c9-account-create-m7dhb"] Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.045886 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndjgw\" (UniqueName: \"kubernetes.io/projected/6a2d8750-d82d-4a23-ac9f-20e4331381e3-kube-api-access-ndjgw\") pod \"placement-43c9-account-create-m7dhb\" (UID: \"6a2d8750-d82d-4a23-ac9f-20e4331381e3\") " pod="openstack/placement-43c9-account-create-m7dhb" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.147076 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndjgw\" (UniqueName: \"kubernetes.io/projected/6a2d8750-d82d-4a23-ac9f-20e4331381e3-kube-api-access-ndjgw\") pod \"placement-43c9-account-create-m7dhb\" (UID: \"6a2d8750-d82d-4a23-ac9f-20e4331381e3\") " pod="openstack/placement-43c9-account-create-m7dhb" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.164514 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndjgw\" (UniqueName: \"kubernetes.io/projected/6a2d8750-d82d-4a23-ac9f-20e4331381e3-kube-api-access-ndjgw\") pod \"placement-43c9-account-create-m7dhb\" (UID: \"6a2d8750-d82d-4a23-ac9f-20e4331381e3\") " pod="openstack/placement-43c9-account-create-m7dhb" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.291441 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-q5v9v" podUID="603a8fc2-2eac-47b7-8e97-cd117033a313" containerName="ovn-controller" probeResult="failure" output=< Oct 07 14:02:02 crc kubenswrapper[4959]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 07 14:02:02 crc kubenswrapper[4959]: > Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.303647 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-43c9-account-create-m7dhb" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.384761 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-c1a2-account-create-xbdrw"] Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.400053 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-fjrkz"] Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.401328 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-fjrkz" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.402553 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-v6cb5" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.404917 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-bhw4d" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.405519 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.407066 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-fjrkz"] Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.447001 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-v6cb5" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.553605 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/024f57ee-b5d4-4f45-b447-5410cf1722ec-combined-ca-bundle\") pod \"glance-db-sync-fjrkz\" (UID: \"024f57ee-b5d4-4f45-b447-5410cf1722ec\") " pod="openstack/glance-db-sync-fjrkz" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.553662 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/024f57ee-b5d4-4f45-b447-5410cf1722ec-db-sync-config-data\") pod \"glance-db-sync-fjrkz\" (UID: \"024f57ee-b5d4-4f45-b447-5410cf1722ec\") " pod="openstack/glance-db-sync-fjrkz" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.553769 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7sq2\" (UniqueName: \"kubernetes.io/projected/024f57ee-b5d4-4f45-b447-5410cf1722ec-kube-api-access-k7sq2\") pod \"glance-db-sync-fjrkz\" (UID: \"024f57ee-b5d4-4f45-b447-5410cf1722ec\") " pod="openstack/glance-db-sync-fjrkz" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.553790 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/024f57ee-b5d4-4f45-b447-5410cf1722ec-config-data\") pod \"glance-db-sync-fjrkz\" (UID: \"024f57ee-b5d4-4f45-b447-5410cf1722ec\") " pod="openstack/glance-db-sync-fjrkz" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.656245 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7sq2\" (UniqueName: \"kubernetes.io/projected/024f57ee-b5d4-4f45-b447-5410cf1722ec-kube-api-access-k7sq2\") pod \"glance-db-sync-fjrkz\" (UID: \"024f57ee-b5d4-4f45-b447-5410cf1722ec\") " pod="openstack/glance-db-sync-fjrkz" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.656307 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/024f57ee-b5d4-4f45-b447-5410cf1722ec-config-data\") pod \"glance-db-sync-fjrkz\" (UID: \"024f57ee-b5d4-4f45-b447-5410cf1722ec\") " pod="openstack/glance-db-sync-fjrkz" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.656365 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/024f57ee-b5d4-4f45-b447-5410cf1722ec-combined-ca-bundle\") pod \"glance-db-sync-fjrkz\" (UID: \"024f57ee-b5d4-4f45-b447-5410cf1722ec\") " pod="openstack/glance-db-sync-fjrkz" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.656409 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/024f57ee-b5d4-4f45-b447-5410cf1722ec-db-sync-config-data\") pod \"glance-db-sync-fjrkz\" (UID: \"024f57ee-b5d4-4f45-b447-5410cf1722ec\") " pod="openstack/glance-db-sync-fjrkz" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.668725 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/024f57ee-b5d4-4f45-b447-5410cf1722ec-combined-ca-bundle\") pod \"glance-db-sync-fjrkz\" (UID: \"024f57ee-b5d4-4f45-b447-5410cf1722ec\") " pod="openstack/glance-db-sync-fjrkz" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.671997 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/024f57ee-b5d4-4f45-b447-5410cf1722ec-db-sync-config-data\") pod \"glance-db-sync-fjrkz\" (UID: \"024f57ee-b5d4-4f45-b447-5410cf1722ec\") " pod="openstack/glance-db-sync-fjrkz" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.683165 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/024f57ee-b5d4-4f45-b447-5410cf1722ec-config-data\") pod \"glance-db-sync-fjrkz\" (UID: \"024f57ee-b5d4-4f45-b447-5410cf1722ec\") " pod="openstack/glance-db-sync-fjrkz" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.696650 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-q5v9v-config-dczfb"] Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.706298 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-q5v9v-config-dczfb" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.710294 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.746848 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-q5v9v-config-dczfb"] Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.750042 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7sq2\" (UniqueName: \"kubernetes.io/projected/024f57ee-b5d4-4f45-b447-5410cf1722ec-kube-api-access-k7sq2\") pod \"glance-db-sync-fjrkz\" (UID: \"024f57ee-b5d4-4f45-b447-5410cf1722ec\") " pod="openstack/glance-db-sync-fjrkz" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.866989 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/580bf374-95d8-45cf-abd8-7077b68847ae-var-run-ovn\") pod \"ovn-controller-q5v9v-config-dczfb\" (UID: \"580bf374-95d8-45cf-abd8-7077b68847ae\") " pod="openstack/ovn-controller-q5v9v-config-dczfb" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.867032 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/580bf374-95d8-45cf-abd8-7077b68847ae-var-log-ovn\") pod \"ovn-controller-q5v9v-config-dczfb\" (UID: \"580bf374-95d8-45cf-abd8-7077b68847ae\") " pod="openstack/ovn-controller-q5v9v-config-dczfb" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.867057 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/580bf374-95d8-45cf-abd8-7077b68847ae-additional-scripts\") pod \"ovn-controller-q5v9v-config-dczfb\" (UID: \"580bf374-95d8-45cf-abd8-7077b68847ae\") " pod="openstack/ovn-controller-q5v9v-config-dczfb" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.867093 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/580bf374-95d8-45cf-abd8-7077b68847ae-var-run\") pod \"ovn-controller-q5v9v-config-dczfb\" (UID: \"580bf374-95d8-45cf-abd8-7077b68847ae\") " pod="openstack/ovn-controller-q5v9v-config-dczfb" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.867129 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zhcj\" (UniqueName: \"kubernetes.io/projected/580bf374-95d8-45cf-abd8-7077b68847ae-kube-api-access-9zhcj\") pod \"ovn-controller-q5v9v-config-dczfb\" (UID: \"580bf374-95d8-45cf-abd8-7077b68847ae\") " pod="openstack/ovn-controller-q5v9v-config-dczfb" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.867153 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/580bf374-95d8-45cf-abd8-7077b68847ae-scripts\") pod \"ovn-controller-q5v9v-config-dczfb\" (UID: \"580bf374-95d8-45cf-abd8-7077b68847ae\") " pod="openstack/ovn-controller-q5v9v-config-dczfb" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.896320 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-43c9-account-create-m7dhb"] Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.968615 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/580bf374-95d8-45cf-abd8-7077b68847ae-var-run\") pod \"ovn-controller-q5v9v-config-dczfb\" (UID: \"580bf374-95d8-45cf-abd8-7077b68847ae\") " pod="openstack/ovn-controller-q5v9v-config-dczfb" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.968672 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zhcj\" (UniqueName: \"kubernetes.io/projected/580bf374-95d8-45cf-abd8-7077b68847ae-kube-api-access-9zhcj\") pod \"ovn-controller-q5v9v-config-dczfb\" (UID: \"580bf374-95d8-45cf-abd8-7077b68847ae\") " pod="openstack/ovn-controller-q5v9v-config-dczfb" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.968699 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/580bf374-95d8-45cf-abd8-7077b68847ae-scripts\") pod \"ovn-controller-q5v9v-config-dczfb\" (UID: \"580bf374-95d8-45cf-abd8-7077b68847ae\") " pod="openstack/ovn-controller-q5v9v-config-dczfb" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.968833 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/580bf374-95d8-45cf-abd8-7077b68847ae-var-run-ovn\") pod \"ovn-controller-q5v9v-config-dczfb\" (UID: \"580bf374-95d8-45cf-abd8-7077b68847ae\") " pod="openstack/ovn-controller-q5v9v-config-dczfb" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.968855 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/580bf374-95d8-45cf-abd8-7077b68847ae-var-log-ovn\") pod \"ovn-controller-q5v9v-config-dczfb\" (UID: \"580bf374-95d8-45cf-abd8-7077b68847ae\") " pod="openstack/ovn-controller-q5v9v-config-dczfb" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.968872 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/580bf374-95d8-45cf-abd8-7077b68847ae-additional-scripts\") pod \"ovn-controller-q5v9v-config-dczfb\" (UID: \"580bf374-95d8-45cf-abd8-7077b68847ae\") " pod="openstack/ovn-controller-q5v9v-config-dczfb" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.968973 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/580bf374-95d8-45cf-abd8-7077b68847ae-var-run\") pod \"ovn-controller-q5v9v-config-dczfb\" (UID: \"580bf374-95d8-45cf-abd8-7077b68847ae\") " pod="openstack/ovn-controller-q5v9v-config-dczfb" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.969053 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/580bf374-95d8-45cf-abd8-7077b68847ae-var-run-ovn\") pod \"ovn-controller-q5v9v-config-dczfb\" (UID: \"580bf374-95d8-45cf-abd8-7077b68847ae\") " pod="openstack/ovn-controller-q5v9v-config-dczfb" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.969052 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/580bf374-95d8-45cf-abd8-7077b68847ae-var-log-ovn\") pod \"ovn-controller-q5v9v-config-dczfb\" (UID: \"580bf374-95d8-45cf-abd8-7077b68847ae\") " pod="openstack/ovn-controller-q5v9v-config-dczfb" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.969740 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/580bf374-95d8-45cf-abd8-7077b68847ae-additional-scripts\") pod \"ovn-controller-q5v9v-config-dczfb\" (UID: \"580bf374-95d8-45cf-abd8-7077b68847ae\") " pod="openstack/ovn-controller-q5v9v-config-dczfb" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.970846 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/580bf374-95d8-45cf-abd8-7077b68847ae-scripts\") pod \"ovn-controller-q5v9v-config-dczfb\" (UID: \"580bf374-95d8-45cf-abd8-7077b68847ae\") " pod="openstack/ovn-controller-q5v9v-config-dczfb" Oct 07 14:02:02 crc kubenswrapper[4959]: I1007 14:02:02.984868 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zhcj\" (UniqueName: \"kubernetes.io/projected/580bf374-95d8-45cf-abd8-7077b68847ae-kube-api-access-9zhcj\") pod \"ovn-controller-q5v9v-config-dczfb\" (UID: \"580bf374-95d8-45cf-abd8-7077b68847ae\") " pod="openstack/ovn-controller-q5v9v-config-dczfb" Oct 07 14:02:03 crc kubenswrapper[4959]: I1007 14:02:03.026457 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-fjrkz" Oct 07 14:02:03 crc kubenswrapper[4959]: I1007 14:02:03.054627 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-q5v9v-config-dczfb" Oct 07 14:02:03 crc kubenswrapper[4959]: I1007 14:02:03.168558 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 07 14:02:03 crc kubenswrapper[4959]: I1007 14:02:03.263857 4959 generic.go:334] "Generic (PLEG): container finished" podID="ccf90649-6706-47f5-a899-f8674a14efcd" containerID="1d7299eb35bfd1ec5e01a5d11f283b0e2b84c38ad8c0ca17eb953750bc8426a5" exitCode=0 Oct 07 14:02:03 crc kubenswrapper[4959]: I1007 14:02:03.263935 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c1a2-account-create-xbdrw" event={"ID":"ccf90649-6706-47f5-a899-f8674a14efcd","Type":"ContainerDied","Data":"1d7299eb35bfd1ec5e01a5d11f283b0e2b84c38ad8c0ca17eb953750bc8426a5"} Oct 07 14:02:03 crc kubenswrapper[4959]: I1007 14:02:03.263991 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c1a2-account-create-xbdrw" event={"ID":"ccf90649-6706-47f5-a899-f8674a14efcd","Type":"ContainerStarted","Data":"215340f901d51e4056fe48894d90ffe5ac32a1a45c540261c22536190cd44d27"} Oct 07 14:02:03 crc kubenswrapper[4959]: I1007 14:02:03.265546 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-43c9-account-create-m7dhb" event={"ID":"6a2d8750-d82d-4a23-ac9f-20e4331381e3","Type":"ContainerStarted","Data":"b7bc97cf6fad5bb90fe097c0d43b1ea587884035f61d602f2e3f105ff5170c97"} Oct 07 14:02:03 crc kubenswrapper[4959]: W1007 14:02:03.537321 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod580bf374_95d8_45cf_abd8_7077b68847ae.slice/crio-7456a7fc34ad306c2dfc1b6edb171f27a2e51b74f9bbd6cefe3520ba0e2aa457 WatchSource:0}: Error finding container 7456a7fc34ad306c2dfc1b6edb171f27a2e51b74f9bbd6cefe3520ba0e2aa457: Status 404 returned error can't find the container with id 7456a7fc34ad306c2dfc1b6edb171f27a2e51b74f9bbd6cefe3520ba0e2aa457 Oct 07 14:02:03 crc kubenswrapper[4959]: I1007 14:02:03.537431 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-q5v9v-config-dczfb"] Oct 07 14:02:03 crc kubenswrapper[4959]: I1007 14:02:03.707926 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-fjrkz"] Oct 07 14:02:03 crc kubenswrapper[4959]: W1007 14:02:03.719905 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod024f57ee_b5d4_4f45_b447_5410cf1722ec.slice/crio-09880e9a444d85aca439b39c442908a1c3f79e08dfa55f047365308af4869ab2 WatchSource:0}: Error finding container 09880e9a444d85aca439b39c442908a1c3f79e08dfa55f047365308af4869ab2: Status 404 returned error can't find the container with id 09880e9a444d85aca439b39c442908a1c3f79e08dfa55f047365308af4869ab2 Oct 07 14:02:04 crc kubenswrapper[4959]: I1007 14:02:04.272370 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-fjrkz" event={"ID":"024f57ee-b5d4-4f45-b447-5410cf1722ec","Type":"ContainerStarted","Data":"09880e9a444d85aca439b39c442908a1c3f79e08dfa55f047365308af4869ab2"} Oct 07 14:02:04 crc kubenswrapper[4959]: I1007 14:02:04.273915 4959 generic.go:334] "Generic (PLEG): container finished" podID="6a2d8750-d82d-4a23-ac9f-20e4331381e3" containerID="0138d038f76b86146c8dbb8e948f7c884f6720f31483a9b2900a96a1ec074d4a" exitCode=0 Oct 07 14:02:04 crc kubenswrapper[4959]: I1007 14:02:04.273957 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-43c9-account-create-m7dhb" event={"ID":"6a2d8750-d82d-4a23-ac9f-20e4331381e3","Type":"ContainerDied","Data":"0138d038f76b86146c8dbb8e948f7c884f6720f31483a9b2900a96a1ec074d4a"} Oct 07 14:02:04 crc kubenswrapper[4959]: I1007 14:02:04.275916 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-q5v9v-config-dczfb" event={"ID":"580bf374-95d8-45cf-abd8-7077b68847ae","Type":"ContainerStarted","Data":"c0ad2f18829f8f0a34b2c312218f32a2d59a1eb1a2c1b95f5eb697e0b023c290"} Oct 07 14:02:04 crc kubenswrapper[4959]: I1007 14:02:04.275973 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-q5v9v-config-dczfb" event={"ID":"580bf374-95d8-45cf-abd8-7077b68847ae","Type":"ContainerStarted","Data":"7456a7fc34ad306c2dfc1b6edb171f27a2e51b74f9bbd6cefe3520ba0e2aa457"} Oct 07 14:02:04 crc kubenswrapper[4959]: I1007 14:02:04.316215 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-q5v9v-config-dczfb" podStartSLOduration=2.316195392 podStartE2EDuration="2.316195392s" podCreationTimestamp="2025-10-07 14:02:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:02:04.305990617 +0000 UTC m=+1006.389394932" watchObservedRunningTime="2025-10-07 14:02:04.316195392 +0000 UTC m=+1006.399599717" Oct 07 14:02:04 crc kubenswrapper[4959]: I1007 14:02:04.613305 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c1a2-account-create-xbdrw" Oct 07 14:02:04 crc kubenswrapper[4959]: I1007 14:02:04.701804 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7l95\" (UniqueName: \"kubernetes.io/projected/ccf90649-6706-47f5-a899-f8674a14efcd-kube-api-access-p7l95\") pod \"ccf90649-6706-47f5-a899-f8674a14efcd\" (UID: \"ccf90649-6706-47f5-a899-f8674a14efcd\") " Oct 07 14:02:04 crc kubenswrapper[4959]: I1007 14:02:04.713505 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccf90649-6706-47f5-a899-f8674a14efcd-kube-api-access-p7l95" (OuterVolumeSpecName: "kube-api-access-p7l95") pod "ccf90649-6706-47f5-a899-f8674a14efcd" (UID: "ccf90649-6706-47f5-a899-f8674a14efcd"). InnerVolumeSpecName "kube-api-access-p7l95". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:02:04 crc kubenswrapper[4959]: I1007 14:02:04.804229 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7l95\" (UniqueName: \"kubernetes.io/projected/ccf90649-6706-47f5-a899-f8674a14efcd-kube-api-access-p7l95\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:05 crc kubenswrapper[4959]: I1007 14:02:05.289862 4959 generic.go:334] "Generic (PLEG): container finished" podID="580bf374-95d8-45cf-abd8-7077b68847ae" containerID="c0ad2f18829f8f0a34b2c312218f32a2d59a1eb1a2c1b95f5eb697e0b023c290" exitCode=0 Oct 07 14:02:05 crc kubenswrapper[4959]: I1007 14:02:05.289975 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-q5v9v-config-dczfb" event={"ID":"580bf374-95d8-45cf-abd8-7077b68847ae","Type":"ContainerDied","Data":"c0ad2f18829f8f0a34b2c312218f32a2d59a1eb1a2c1b95f5eb697e0b023c290"} Oct 07 14:02:05 crc kubenswrapper[4959]: I1007 14:02:05.293869 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c1a2-account-create-xbdrw" Oct 07 14:02:05 crc kubenswrapper[4959]: I1007 14:02:05.293910 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c1a2-account-create-xbdrw" event={"ID":"ccf90649-6706-47f5-a899-f8674a14efcd","Type":"ContainerDied","Data":"215340f901d51e4056fe48894d90ffe5ac32a1a45c540261c22536190cd44d27"} Oct 07 14:02:05 crc kubenswrapper[4959]: I1007 14:02:05.293933 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="215340f901d51e4056fe48894d90ffe5ac32a1a45c540261c22536190cd44d27" Oct 07 14:02:05 crc kubenswrapper[4959]: I1007 14:02:05.649728 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-43c9-account-create-m7dhb" Oct 07 14:02:05 crc kubenswrapper[4959]: I1007 14:02:05.822618 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndjgw\" (UniqueName: \"kubernetes.io/projected/6a2d8750-d82d-4a23-ac9f-20e4331381e3-kube-api-access-ndjgw\") pod \"6a2d8750-d82d-4a23-ac9f-20e4331381e3\" (UID: \"6a2d8750-d82d-4a23-ac9f-20e4331381e3\") " Oct 07 14:02:05 crc kubenswrapper[4959]: I1007 14:02:05.826284 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a2d8750-d82d-4a23-ac9f-20e4331381e3-kube-api-access-ndjgw" (OuterVolumeSpecName: "kube-api-access-ndjgw") pod "6a2d8750-d82d-4a23-ac9f-20e4331381e3" (UID: "6a2d8750-d82d-4a23-ac9f-20e4331381e3"). InnerVolumeSpecName "kube-api-access-ndjgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:02:05 crc kubenswrapper[4959]: I1007 14:02:05.924205 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndjgw\" (UniqueName: \"kubernetes.io/projected/6a2d8750-d82d-4a23-ac9f-20e4331381e3-kube-api-access-ndjgw\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:06 crc kubenswrapper[4959]: I1007 14:02:06.303413 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-43c9-account-create-m7dhb" Oct 07 14:02:06 crc kubenswrapper[4959]: I1007 14:02:06.303449 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-43c9-account-create-m7dhb" event={"ID":"6a2d8750-d82d-4a23-ac9f-20e4331381e3","Type":"ContainerDied","Data":"b7bc97cf6fad5bb90fe097c0d43b1ea587884035f61d602f2e3f105ff5170c97"} Oct 07 14:02:06 crc kubenswrapper[4959]: I1007 14:02:06.303490 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7bc97cf6fad5bb90fe097c0d43b1ea587884035f61d602f2e3f105ff5170c97" Oct 07 14:02:06 crc kubenswrapper[4959]: I1007 14:02:06.658763 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-q5v9v-config-dczfb" Oct 07 14:02:06 crc kubenswrapper[4959]: I1007 14:02:06.737255 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zhcj\" (UniqueName: \"kubernetes.io/projected/580bf374-95d8-45cf-abd8-7077b68847ae-kube-api-access-9zhcj\") pod \"580bf374-95d8-45cf-abd8-7077b68847ae\" (UID: \"580bf374-95d8-45cf-abd8-7077b68847ae\") " Oct 07 14:02:06 crc kubenswrapper[4959]: I1007 14:02:06.737711 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/580bf374-95d8-45cf-abd8-7077b68847ae-var-run-ovn\") pod \"580bf374-95d8-45cf-abd8-7077b68847ae\" (UID: \"580bf374-95d8-45cf-abd8-7077b68847ae\") " Oct 07 14:02:06 crc kubenswrapper[4959]: I1007 14:02:06.737800 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/580bf374-95d8-45cf-abd8-7077b68847ae-scripts\") pod \"580bf374-95d8-45cf-abd8-7077b68847ae\" (UID: \"580bf374-95d8-45cf-abd8-7077b68847ae\") " Oct 07 14:02:06 crc kubenswrapper[4959]: I1007 14:02:06.737828 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/580bf374-95d8-45cf-abd8-7077b68847ae-additional-scripts\") pod \"580bf374-95d8-45cf-abd8-7077b68847ae\" (UID: \"580bf374-95d8-45cf-abd8-7077b68847ae\") " Oct 07 14:02:06 crc kubenswrapper[4959]: I1007 14:02:06.737871 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/580bf374-95d8-45cf-abd8-7077b68847ae-var-log-ovn\") pod \"580bf374-95d8-45cf-abd8-7077b68847ae\" (UID: \"580bf374-95d8-45cf-abd8-7077b68847ae\") " Oct 07 14:02:06 crc kubenswrapper[4959]: I1007 14:02:06.737930 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/580bf374-95d8-45cf-abd8-7077b68847ae-var-run\") pod \"580bf374-95d8-45cf-abd8-7077b68847ae\" (UID: \"580bf374-95d8-45cf-abd8-7077b68847ae\") " Oct 07 14:02:06 crc kubenswrapper[4959]: I1007 14:02:06.737784 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/580bf374-95d8-45cf-abd8-7077b68847ae-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "580bf374-95d8-45cf-abd8-7077b68847ae" (UID: "580bf374-95d8-45cf-abd8-7077b68847ae"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 14:02:06 crc kubenswrapper[4959]: I1007 14:02:06.738279 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/580bf374-95d8-45cf-abd8-7077b68847ae-var-run" (OuterVolumeSpecName: "var-run") pod "580bf374-95d8-45cf-abd8-7077b68847ae" (UID: "580bf374-95d8-45cf-abd8-7077b68847ae"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 14:02:06 crc kubenswrapper[4959]: I1007 14:02:06.738837 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/580bf374-95d8-45cf-abd8-7077b68847ae-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "580bf374-95d8-45cf-abd8-7077b68847ae" (UID: "580bf374-95d8-45cf-abd8-7077b68847ae"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:02:06 crc kubenswrapper[4959]: I1007 14:02:06.738869 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/580bf374-95d8-45cf-abd8-7077b68847ae-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "580bf374-95d8-45cf-abd8-7077b68847ae" (UID: "580bf374-95d8-45cf-abd8-7077b68847ae"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 14:02:06 crc kubenswrapper[4959]: I1007 14:02:06.738854 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/580bf374-95d8-45cf-abd8-7077b68847ae-scripts" (OuterVolumeSpecName: "scripts") pod "580bf374-95d8-45cf-abd8-7077b68847ae" (UID: "580bf374-95d8-45cf-abd8-7077b68847ae"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:02:06 crc kubenswrapper[4959]: I1007 14:02:06.742945 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/580bf374-95d8-45cf-abd8-7077b68847ae-kube-api-access-9zhcj" (OuterVolumeSpecName: "kube-api-access-9zhcj") pod "580bf374-95d8-45cf-abd8-7077b68847ae" (UID: "580bf374-95d8-45cf-abd8-7077b68847ae"). InnerVolumeSpecName "kube-api-access-9zhcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:02:06 crc kubenswrapper[4959]: I1007 14:02:06.839614 4959 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/580bf374-95d8-45cf-abd8-7077b68847ae-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:06 crc kubenswrapper[4959]: I1007 14:02:06.839660 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/580bf374-95d8-45cf-abd8-7077b68847ae-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:06 crc kubenswrapper[4959]: I1007 14:02:06.839673 4959 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/580bf374-95d8-45cf-abd8-7077b68847ae-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:06 crc kubenswrapper[4959]: I1007 14:02:06.839686 4959 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/580bf374-95d8-45cf-abd8-7077b68847ae-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:06 crc kubenswrapper[4959]: I1007 14:02:06.839696 4959 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/580bf374-95d8-45cf-abd8-7077b68847ae-var-run\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:06 crc kubenswrapper[4959]: I1007 14:02:06.839708 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zhcj\" (UniqueName: \"kubernetes.io/projected/580bf374-95d8-45cf-abd8-7077b68847ae-kube-api-access-9zhcj\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.295690 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-q5v9v" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.323264 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-q5v9v-config-dczfb" event={"ID":"580bf374-95d8-45cf-abd8-7077b68847ae","Type":"ContainerDied","Data":"7456a7fc34ad306c2dfc1b6edb171f27a2e51b74f9bbd6cefe3520ba0e2aa457"} Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.323310 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7456a7fc34ad306c2dfc1b6edb171f27a2e51b74f9bbd6cefe3520ba0e2aa457" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.323370 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-q5v9v-config-dczfb" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.394668 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-q5v9v-config-dczfb"] Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.406055 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-q5v9v-config-dczfb"] Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.494890 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-q5v9v-config-x7759"] Oct 07 14:02:07 crc kubenswrapper[4959]: E1007 14:02:07.496621 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccf90649-6706-47f5-a899-f8674a14efcd" containerName="mariadb-account-create" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.496665 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccf90649-6706-47f5-a899-f8674a14efcd" containerName="mariadb-account-create" Oct 07 14:02:07 crc kubenswrapper[4959]: E1007 14:02:07.496681 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="580bf374-95d8-45cf-abd8-7077b68847ae" containerName="ovn-config" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.496689 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="580bf374-95d8-45cf-abd8-7077b68847ae" containerName="ovn-config" Oct 07 14:02:07 crc kubenswrapper[4959]: E1007 14:02:07.496711 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a2d8750-d82d-4a23-ac9f-20e4331381e3" containerName="mariadb-account-create" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.496718 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a2d8750-d82d-4a23-ac9f-20e4331381e3" containerName="mariadb-account-create" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.496893 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccf90649-6706-47f5-a899-f8674a14efcd" containerName="mariadb-account-create" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.496909 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a2d8750-d82d-4a23-ac9f-20e4331381e3" containerName="mariadb-account-create" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.496928 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="580bf374-95d8-45cf-abd8-7077b68847ae" containerName="ovn-config" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.497617 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-q5v9v-config-x7759" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.500751 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.508788 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-q5v9v-config-x7759"] Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.651903 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f677e371-82ab-45eb-b358-a2a3ad838137-var-run-ovn\") pod \"ovn-controller-q5v9v-config-x7759\" (UID: \"f677e371-82ab-45eb-b358-a2a3ad838137\") " pod="openstack/ovn-controller-q5v9v-config-x7759" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.652300 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f677e371-82ab-45eb-b358-a2a3ad838137-var-run\") pod \"ovn-controller-q5v9v-config-x7759\" (UID: \"f677e371-82ab-45eb-b358-a2a3ad838137\") " pod="openstack/ovn-controller-q5v9v-config-x7759" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.652434 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdsbt\" (UniqueName: \"kubernetes.io/projected/f677e371-82ab-45eb-b358-a2a3ad838137-kube-api-access-pdsbt\") pod \"ovn-controller-q5v9v-config-x7759\" (UID: \"f677e371-82ab-45eb-b358-a2a3ad838137\") " pod="openstack/ovn-controller-q5v9v-config-x7759" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.652824 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f677e371-82ab-45eb-b358-a2a3ad838137-additional-scripts\") pod \"ovn-controller-q5v9v-config-x7759\" (UID: \"f677e371-82ab-45eb-b358-a2a3ad838137\") " pod="openstack/ovn-controller-q5v9v-config-x7759" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.652935 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f677e371-82ab-45eb-b358-a2a3ad838137-scripts\") pod \"ovn-controller-q5v9v-config-x7759\" (UID: \"f677e371-82ab-45eb-b358-a2a3ad838137\") " pod="openstack/ovn-controller-q5v9v-config-x7759" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.653330 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f677e371-82ab-45eb-b358-a2a3ad838137-var-log-ovn\") pod \"ovn-controller-q5v9v-config-x7759\" (UID: \"f677e371-82ab-45eb-b358-a2a3ad838137\") " pod="openstack/ovn-controller-q5v9v-config-x7759" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.755519 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f677e371-82ab-45eb-b358-a2a3ad838137-additional-scripts\") pod \"ovn-controller-q5v9v-config-x7759\" (UID: \"f677e371-82ab-45eb-b358-a2a3ad838137\") " pod="openstack/ovn-controller-q5v9v-config-x7759" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.755570 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f677e371-82ab-45eb-b358-a2a3ad838137-scripts\") pod \"ovn-controller-q5v9v-config-x7759\" (UID: \"f677e371-82ab-45eb-b358-a2a3ad838137\") " pod="openstack/ovn-controller-q5v9v-config-x7759" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.755602 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f677e371-82ab-45eb-b358-a2a3ad838137-var-log-ovn\") pod \"ovn-controller-q5v9v-config-x7759\" (UID: \"f677e371-82ab-45eb-b358-a2a3ad838137\") " pod="openstack/ovn-controller-q5v9v-config-x7759" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.755672 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f677e371-82ab-45eb-b358-a2a3ad838137-var-run-ovn\") pod \"ovn-controller-q5v9v-config-x7759\" (UID: \"f677e371-82ab-45eb-b358-a2a3ad838137\") " pod="openstack/ovn-controller-q5v9v-config-x7759" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.755709 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f677e371-82ab-45eb-b358-a2a3ad838137-var-run\") pod \"ovn-controller-q5v9v-config-x7759\" (UID: \"f677e371-82ab-45eb-b358-a2a3ad838137\") " pod="openstack/ovn-controller-q5v9v-config-x7759" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.755738 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdsbt\" (UniqueName: \"kubernetes.io/projected/f677e371-82ab-45eb-b358-a2a3ad838137-kube-api-access-pdsbt\") pod \"ovn-controller-q5v9v-config-x7759\" (UID: \"f677e371-82ab-45eb-b358-a2a3ad838137\") " pod="openstack/ovn-controller-q5v9v-config-x7759" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.756000 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f677e371-82ab-45eb-b358-a2a3ad838137-var-log-ovn\") pod \"ovn-controller-q5v9v-config-x7759\" (UID: \"f677e371-82ab-45eb-b358-a2a3ad838137\") " pod="openstack/ovn-controller-q5v9v-config-x7759" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.756073 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f677e371-82ab-45eb-b358-a2a3ad838137-var-run\") pod \"ovn-controller-q5v9v-config-x7759\" (UID: \"f677e371-82ab-45eb-b358-a2a3ad838137\") " pod="openstack/ovn-controller-q5v9v-config-x7759" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.756078 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f677e371-82ab-45eb-b358-a2a3ad838137-var-run-ovn\") pod \"ovn-controller-q5v9v-config-x7759\" (UID: \"f677e371-82ab-45eb-b358-a2a3ad838137\") " pod="openstack/ovn-controller-q5v9v-config-x7759" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.756451 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f677e371-82ab-45eb-b358-a2a3ad838137-additional-scripts\") pod \"ovn-controller-q5v9v-config-x7759\" (UID: \"f677e371-82ab-45eb-b358-a2a3ad838137\") " pod="openstack/ovn-controller-q5v9v-config-x7759" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.758052 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f677e371-82ab-45eb-b358-a2a3ad838137-scripts\") pod \"ovn-controller-q5v9v-config-x7759\" (UID: \"f677e371-82ab-45eb-b358-a2a3ad838137\") " pod="openstack/ovn-controller-q5v9v-config-x7759" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.775307 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdsbt\" (UniqueName: \"kubernetes.io/projected/f677e371-82ab-45eb-b358-a2a3ad838137-kube-api-access-pdsbt\") pod \"ovn-controller-q5v9v-config-x7759\" (UID: \"f677e371-82ab-45eb-b358-a2a3ad838137\") " pod="openstack/ovn-controller-q5v9v-config-x7759" Oct 07 14:02:07 crc kubenswrapper[4959]: I1007 14:02:07.814890 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-q5v9v-config-x7759" Oct 07 14:02:08 crc kubenswrapper[4959]: I1007 14:02:08.318470 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-q5v9v-config-x7759"] Oct 07 14:02:08 crc kubenswrapper[4959]: I1007 14:02:08.631327 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:02:08 crc kubenswrapper[4959]: I1007 14:02:08.670035 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="580bf374-95d8-45cf-abd8-7077b68847ae" path="/var/lib/kubelet/pods/580bf374-95d8-45cf-abd8-7077b68847ae/volumes" Oct 07 14:02:08 crc kubenswrapper[4959]: I1007 14:02:08.670788 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 07 14:02:09 crc kubenswrapper[4959]: I1007 14:02:09.349332 4959 generic.go:334] "Generic (PLEG): container finished" podID="f677e371-82ab-45eb-b358-a2a3ad838137" containerID="bbbf601c5426d5a56d29c92700a02221d86468a30f2fbe80f508b782d4182e22" exitCode=0 Oct 07 14:02:09 crc kubenswrapper[4959]: I1007 14:02:09.349612 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-q5v9v-config-x7759" event={"ID":"f677e371-82ab-45eb-b358-a2a3ad838137","Type":"ContainerDied","Data":"bbbf601c5426d5a56d29c92700a02221d86468a30f2fbe80f508b782d4182e22"} Oct 07 14:02:09 crc kubenswrapper[4959]: I1007 14:02:09.349637 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-q5v9v-config-x7759" event={"ID":"f677e371-82ab-45eb-b358-a2a3ad838137","Type":"ContainerStarted","Data":"2586a57d00f1d8dd657e3075c864d8d0fa4a3649f496d69461aff67ff7a09db2"} Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.414680 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-qgn9g"] Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.416215 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qgn9g" Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.430198 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-qgn9g"] Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.504310 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2jbb\" (UniqueName: \"kubernetes.io/projected/010fee6b-bceb-4a25-b50f-19ccc1eb79a7-kube-api-access-r2jbb\") pod \"cinder-db-create-qgn9g\" (UID: \"010fee6b-bceb-4a25-b50f-19ccc1eb79a7\") " pod="openstack/cinder-db-create-qgn9g" Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.515435 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-btxq4"] Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.517659 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-btxq4" Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.535028 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-btxq4"] Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.607001 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2jbb\" (UniqueName: \"kubernetes.io/projected/010fee6b-bceb-4a25-b50f-19ccc1eb79a7-kube-api-access-r2jbb\") pod \"cinder-db-create-qgn9g\" (UID: \"010fee6b-bceb-4a25-b50f-19ccc1eb79a7\") " pod="openstack/cinder-db-create-qgn9g" Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.607179 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vd9bj\" (UniqueName: \"kubernetes.io/projected/2bea0796-b66c-4c44-be6e-840bde657893-kube-api-access-vd9bj\") pod \"barbican-db-create-btxq4\" (UID: \"2bea0796-b66c-4c44-be6e-840bde657893\") " pod="openstack/barbican-db-create-btxq4" Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.620091 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-bpqp9"] Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.621755 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-bpqp9" Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.634355 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2jbb\" (UniqueName: \"kubernetes.io/projected/010fee6b-bceb-4a25-b50f-19ccc1eb79a7-kube-api-access-r2jbb\") pod \"cinder-db-create-qgn9g\" (UID: \"010fee6b-bceb-4a25-b50f-19ccc1eb79a7\") " pod="openstack/cinder-db-create-qgn9g" Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.675720 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-bpqp9"] Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.707063 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-gvngb"] Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.708416 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gvngb" Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.709235 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vd9bj\" (UniqueName: \"kubernetes.io/projected/2bea0796-b66c-4c44-be6e-840bde657893-kube-api-access-vd9bj\") pod \"barbican-db-create-btxq4\" (UID: \"2bea0796-b66c-4c44-be6e-840bde657893\") " pod="openstack/barbican-db-create-btxq4" Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.709346 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48fcl\" (UniqueName: \"kubernetes.io/projected/32df9b0d-1d2c-4724-b6c7-c4902baf8ae0-kube-api-access-48fcl\") pod \"neutron-db-create-bpqp9\" (UID: \"32df9b0d-1d2c-4724-b6c7-c4902baf8ae0\") " pod="openstack/neutron-db-create-bpqp9" Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.711065 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.711767 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.711950 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.712162 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-z8j6b" Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.722681 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-gvngb"] Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.740225 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qgn9g" Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.746069 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vd9bj\" (UniqueName: \"kubernetes.io/projected/2bea0796-b66c-4c44-be6e-840bde657893-kube-api-access-vd9bj\") pod \"barbican-db-create-btxq4\" (UID: \"2bea0796-b66c-4c44-be6e-840bde657893\") " pod="openstack/barbican-db-create-btxq4" Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.810609 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f-config-data\") pod \"keystone-db-sync-gvngb\" (UID: \"c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f\") " pod="openstack/keystone-db-sync-gvngb" Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.810720 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48fcl\" (UniqueName: \"kubernetes.io/projected/32df9b0d-1d2c-4724-b6c7-c4902baf8ae0-kube-api-access-48fcl\") pod \"neutron-db-create-bpqp9\" (UID: \"32df9b0d-1d2c-4724-b6c7-c4902baf8ae0\") " pod="openstack/neutron-db-create-bpqp9" Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.810750 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f-combined-ca-bundle\") pod \"keystone-db-sync-gvngb\" (UID: \"c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f\") " pod="openstack/keystone-db-sync-gvngb" Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.810785 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxv2b\" (UniqueName: \"kubernetes.io/projected/c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f-kube-api-access-lxv2b\") pod \"keystone-db-sync-gvngb\" (UID: \"c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f\") " pod="openstack/keystone-db-sync-gvngb" Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.834009 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48fcl\" (UniqueName: \"kubernetes.io/projected/32df9b0d-1d2c-4724-b6c7-c4902baf8ae0-kube-api-access-48fcl\") pod \"neutron-db-create-bpqp9\" (UID: \"32df9b0d-1d2c-4724-b6c7-c4902baf8ae0\") " pod="openstack/neutron-db-create-bpqp9" Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.857354 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-btxq4" Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.912896 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f-combined-ca-bundle\") pod \"keystone-db-sync-gvngb\" (UID: \"c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f\") " pod="openstack/keystone-db-sync-gvngb" Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.912955 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxv2b\" (UniqueName: \"kubernetes.io/projected/c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f-kube-api-access-lxv2b\") pod \"keystone-db-sync-gvngb\" (UID: \"c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f\") " pod="openstack/keystone-db-sync-gvngb" Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.912996 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f-config-data\") pod \"keystone-db-sync-gvngb\" (UID: \"c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f\") " pod="openstack/keystone-db-sync-gvngb" Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.916261 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f-combined-ca-bundle\") pod \"keystone-db-sync-gvngb\" (UID: \"c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f\") " pod="openstack/keystone-db-sync-gvngb" Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.916537 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f-config-data\") pod \"keystone-db-sync-gvngb\" (UID: \"c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f\") " pod="openstack/keystone-db-sync-gvngb" Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.930019 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxv2b\" (UniqueName: \"kubernetes.io/projected/c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f-kube-api-access-lxv2b\") pod \"keystone-db-sync-gvngb\" (UID: \"c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f\") " pod="openstack/keystone-db-sync-gvngb" Oct 07 14:02:10 crc kubenswrapper[4959]: I1007 14:02:10.980448 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-bpqp9" Oct 07 14:02:11 crc kubenswrapper[4959]: I1007 14:02:11.028622 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gvngb" Oct 07 14:02:15 crc kubenswrapper[4959]: I1007 14:02:15.358686 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-q5v9v-config-x7759" Oct 07 14:02:15 crc kubenswrapper[4959]: I1007 14:02:15.444374 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-q5v9v-config-x7759" event={"ID":"f677e371-82ab-45eb-b358-a2a3ad838137","Type":"ContainerDied","Data":"2586a57d00f1d8dd657e3075c864d8d0fa4a3649f496d69461aff67ff7a09db2"} Oct 07 14:02:15 crc kubenswrapper[4959]: I1007 14:02:15.444664 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2586a57d00f1d8dd657e3075c864d8d0fa4a3649f496d69461aff67ff7a09db2" Oct 07 14:02:15 crc kubenswrapper[4959]: I1007 14:02:15.444755 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-q5v9v-config-x7759" Oct 07 14:02:15 crc kubenswrapper[4959]: I1007 14:02:15.492547 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f677e371-82ab-45eb-b358-a2a3ad838137-additional-scripts\") pod \"f677e371-82ab-45eb-b358-a2a3ad838137\" (UID: \"f677e371-82ab-45eb-b358-a2a3ad838137\") " Oct 07 14:02:15 crc kubenswrapper[4959]: I1007 14:02:15.492638 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f677e371-82ab-45eb-b358-a2a3ad838137-var-run\") pod \"f677e371-82ab-45eb-b358-a2a3ad838137\" (UID: \"f677e371-82ab-45eb-b358-a2a3ad838137\") " Oct 07 14:02:15 crc kubenswrapper[4959]: I1007 14:02:15.492657 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f677e371-82ab-45eb-b358-a2a3ad838137-var-log-ovn\") pod \"f677e371-82ab-45eb-b358-a2a3ad838137\" (UID: \"f677e371-82ab-45eb-b358-a2a3ad838137\") " Oct 07 14:02:15 crc kubenswrapper[4959]: I1007 14:02:15.492679 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdsbt\" (UniqueName: \"kubernetes.io/projected/f677e371-82ab-45eb-b358-a2a3ad838137-kube-api-access-pdsbt\") pod \"f677e371-82ab-45eb-b358-a2a3ad838137\" (UID: \"f677e371-82ab-45eb-b358-a2a3ad838137\") " Oct 07 14:02:15 crc kubenswrapper[4959]: I1007 14:02:15.492702 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f677e371-82ab-45eb-b358-a2a3ad838137-var-run-ovn\") pod \"f677e371-82ab-45eb-b358-a2a3ad838137\" (UID: \"f677e371-82ab-45eb-b358-a2a3ad838137\") " Oct 07 14:02:15 crc kubenswrapper[4959]: I1007 14:02:15.492727 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f677e371-82ab-45eb-b358-a2a3ad838137-var-run" (OuterVolumeSpecName: "var-run") pod "f677e371-82ab-45eb-b358-a2a3ad838137" (UID: "f677e371-82ab-45eb-b358-a2a3ad838137"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 14:02:15 crc kubenswrapper[4959]: I1007 14:02:15.492745 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f677e371-82ab-45eb-b358-a2a3ad838137-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "f677e371-82ab-45eb-b358-a2a3ad838137" (UID: "f677e371-82ab-45eb-b358-a2a3ad838137"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 14:02:15 crc kubenswrapper[4959]: I1007 14:02:15.492841 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f677e371-82ab-45eb-b358-a2a3ad838137-scripts\") pod \"f677e371-82ab-45eb-b358-a2a3ad838137\" (UID: \"f677e371-82ab-45eb-b358-a2a3ad838137\") " Oct 07 14:02:15 crc kubenswrapper[4959]: I1007 14:02:15.492860 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f677e371-82ab-45eb-b358-a2a3ad838137-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "f677e371-82ab-45eb-b358-a2a3ad838137" (UID: "f677e371-82ab-45eb-b358-a2a3ad838137"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 14:02:15 crc kubenswrapper[4959]: I1007 14:02:15.493134 4959 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f677e371-82ab-45eb-b358-a2a3ad838137-var-run\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:15 crc kubenswrapper[4959]: I1007 14:02:15.493145 4959 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f677e371-82ab-45eb-b358-a2a3ad838137-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:15 crc kubenswrapper[4959]: I1007 14:02:15.493155 4959 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f677e371-82ab-45eb-b358-a2a3ad838137-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:15 crc kubenswrapper[4959]: I1007 14:02:15.493643 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f677e371-82ab-45eb-b358-a2a3ad838137-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "f677e371-82ab-45eb-b358-a2a3ad838137" (UID: "f677e371-82ab-45eb-b358-a2a3ad838137"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:02:15 crc kubenswrapper[4959]: I1007 14:02:15.493862 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f677e371-82ab-45eb-b358-a2a3ad838137-scripts" (OuterVolumeSpecName: "scripts") pod "f677e371-82ab-45eb-b358-a2a3ad838137" (UID: "f677e371-82ab-45eb-b358-a2a3ad838137"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:02:15 crc kubenswrapper[4959]: I1007 14:02:15.499508 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f677e371-82ab-45eb-b358-a2a3ad838137-kube-api-access-pdsbt" (OuterVolumeSpecName: "kube-api-access-pdsbt") pod "f677e371-82ab-45eb-b358-a2a3ad838137" (UID: "f677e371-82ab-45eb-b358-a2a3ad838137"). InnerVolumeSpecName "kube-api-access-pdsbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:02:15 crc kubenswrapper[4959]: I1007 14:02:15.506262 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-gvngb"] Oct 07 14:02:15 crc kubenswrapper[4959]: I1007 14:02:15.596065 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f677e371-82ab-45eb-b358-a2a3ad838137-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:15 crc kubenswrapper[4959]: I1007 14:02:15.596091 4959 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f677e371-82ab-45eb-b358-a2a3ad838137-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:15 crc kubenswrapper[4959]: I1007 14:02:15.596115 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdsbt\" (UniqueName: \"kubernetes.io/projected/f677e371-82ab-45eb-b358-a2a3ad838137-kube-api-access-pdsbt\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:15 crc kubenswrapper[4959]: I1007 14:02:15.807275 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-qgn9g"] Oct 07 14:02:15 crc kubenswrapper[4959]: W1007 14:02:15.810297 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod010fee6b_bceb_4a25_b50f_19ccc1eb79a7.slice/crio-fb700398daebae30300b22b1e5c4abd82425bcf5c1ba67555c2dbade6520349e WatchSource:0}: Error finding container fb700398daebae30300b22b1e5c4abd82425bcf5c1ba67555c2dbade6520349e: Status 404 returned error can't find the container with id fb700398daebae30300b22b1e5c4abd82425bcf5c1ba67555c2dbade6520349e Oct 07 14:02:15 crc kubenswrapper[4959]: I1007 14:02:15.871977 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-btxq4"] Oct 07 14:02:15 crc kubenswrapper[4959]: I1007 14:02:15.878389 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-bpqp9"] Oct 07 14:02:15 crc kubenswrapper[4959]: W1007 14:02:15.880605 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32df9b0d_1d2c_4724_b6c7_c4902baf8ae0.slice/crio-9e35b933ecd6375c473514f409887a2aa1aff101cdf1c3d3988e6bd65c58d334 WatchSource:0}: Error finding container 9e35b933ecd6375c473514f409887a2aa1aff101cdf1c3d3988e6bd65c58d334: Status 404 returned error can't find the container with id 9e35b933ecd6375c473514f409887a2aa1aff101cdf1c3d3988e6bd65c58d334 Oct 07 14:02:16 crc kubenswrapper[4959]: I1007 14:02:16.433806 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-q5v9v-config-x7759"] Oct 07 14:02:16 crc kubenswrapper[4959]: I1007 14:02:16.440208 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-q5v9v-config-x7759"] Oct 07 14:02:16 crc kubenswrapper[4959]: I1007 14:02:16.454313 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gvngb" event={"ID":"c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f","Type":"ContainerStarted","Data":"8cb665795ecb28d760aad58e90773b96ca5d9c24ce631440ad74f7bcdf5f16c7"} Oct 07 14:02:16 crc kubenswrapper[4959]: I1007 14:02:16.455559 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-fjrkz" event={"ID":"024f57ee-b5d4-4f45-b447-5410cf1722ec","Type":"ContainerStarted","Data":"612295e06c3281aebb30eac6545902cd9ab7558de4fad84ee6b9eb0cbdc8d549"} Oct 07 14:02:16 crc kubenswrapper[4959]: I1007 14:02:16.456804 4959 generic.go:334] "Generic (PLEG): container finished" podID="32df9b0d-1d2c-4724-b6c7-c4902baf8ae0" containerID="9fb6ccbbc199190c97616fe5220ff9a32382a1179e278b4f7caa3d4e2e13f6ec" exitCode=0 Oct 07 14:02:16 crc kubenswrapper[4959]: I1007 14:02:16.456838 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-bpqp9" event={"ID":"32df9b0d-1d2c-4724-b6c7-c4902baf8ae0","Type":"ContainerDied","Data":"9fb6ccbbc199190c97616fe5220ff9a32382a1179e278b4f7caa3d4e2e13f6ec"} Oct 07 14:02:16 crc kubenswrapper[4959]: I1007 14:02:16.456867 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-bpqp9" event={"ID":"32df9b0d-1d2c-4724-b6c7-c4902baf8ae0","Type":"ContainerStarted","Data":"9e35b933ecd6375c473514f409887a2aa1aff101cdf1c3d3988e6bd65c58d334"} Oct 07 14:02:16 crc kubenswrapper[4959]: I1007 14:02:16.458252 4959 generic.go:334] "Generic (PLEG): container finished" podID="010fee6b-bceb-4a25-b50f-19ccc1eb79a7" containerID="e5664b833d564c90e189d17aa8f318a3e302fb3e3811e034b2ade020fc05e777" exitCode=0 Oct 07 14:02:16 crc kubenswrapper[4959]: I1007 14:02:16.458325 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qgn9g" event={"ID":"010fee6b-bceb-4a25-b50f-19ccc1eb79a7","Type":"ContainerDied","Data":"e5664b833d564c90e189d17aa8f318a3e302fb3e3811e034b2ade020fc05e777"} Oct 07 14:02:16 crc kubenswrapper[4959]: I1007 14:02:16.458351 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qgn9g" event={"ID":"010fee6b-bceb-4a25-b50f-19ccc1eb79a7","Type":"ContainerStarted","Data":"fb700398daebae30300b22b1e5c4abd82425bcf5c1ba67555c2dbade6520349e"} Oct 07 14:02:16 crc kubenswrapper[4959]: I1007 14:02:16.459420 4959 generic.go:334] "Generic (PLEG): container finished" podID="2bea0796-b66c-4c44-be6e-840bde657893" containerID="da70dc6d97f7d723f1e2cd16678e421ed5399ebcb0fb5cce68c6f17e94915a1e" exitCode=0 Oct 07 14:02:16 crc kubenswrapper[4959]: I1007 14:02:16.459442 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-btxq4" event={"ID":"2bea0796-b66c-4c44-be6e-840bde657893","Type":"ContainerDied","Data":"da70dc6d97f7d723f1e2cd16678e421ed5399ebcb0fb5cce68c6f17e94915a1e"} Oct 07 14:02:16 crc kubenswrapper[4959]: I1007 14:02:16.459466 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-btxq4" event={"ID":"2bea0796-b66c-4c44-be6e-840bde657893","Type":"ContainerStarted","Data":"098d6ebdf347d3298668b83253b4529e7022d2be72c04ccd9d0d8661d4e2f002"} Oct 07 14:02:16 crc kubenswrapper[4959]: I1007 14:02:16.477506 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-fjrkz" podStartSLOduration=2.774997949 podStartE2EDuration="14.477468753s" podCreationTimestamp="2025-10-07 14:02:02 +0000 UTC" firstStartedPulling="2025-10-07 14:02:03.7218455 +0000 UTC m=+1005.805249825" lastFinishedPulling="2025-10-07 14:02:15.424316304 +0000 UTC m=+1017.507720629" observedRunningTime="2025-10-07 14:02:16.470834647 +0000 UTC m=+1018.554238972" watchObservedRunningTime="2025-10-07 14:02:16.477468753 +0000 UTC m=+1018.560873088" Oct 07 14:02:16 crc kubenswrapper[4959]: I1007 14:02:16.663757 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f677e371-82ab-45eb-b358-a2a3ad838137" path="/var/lib/kubelet/pods/f677e371-82ab-45eb-b358-a2a3ad838137/volumes" Oct 07 14:02:20 crc kubenswrapper[4959]: I1007 14:02:20.036163 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qgn9g" Oct 07 14:02:20 crc kubenswrapper[4959]: I1007 14:02:20.040731 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-btxq4" Oct 07 14:02:20 crc kubenswrapper[4959]: I1007 14:02:20.047979 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-bpqp9" Oct 07 14:02:20 crc kubenswrapper[4959]: I1007 14:02:20.230484 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48fcl\" (UniqueName: \"kubernetes.io/projected/32df9b0d-1d2c-4724-b6c7-c4902baf8ae0-kube-api-access-48fcl\") pod \"32df9b0d-1d2c-4724-b6c7-c4902baf8ae0\" (UID: \"32df9b0d-1d2c-4724-b6c7-c4902baf8ae0\") " Oct 07 14:02:20 crc kubenswrapper[4959]: I1007 14:02:20.230547 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vd9bj\" (UniqueName: \"kubernetes.io/projected/2bea0796-b66c-4c44-be6e-840bde657893-kube-api-access-vd9bj\") pod \"2bea0796-b66c-4c44-be6e-840bde657893\" (UID: \"2bea0796-b66c-4c44-be6e-840bde657893\") " Oct 07 14:02:20 crc kubenswrapper[4959]: I1007 14:02:20.230691 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2jbb\" (UniqueName: \"kubernetes.io/projected/010fee6b-bceb-4a25-b50f-19ccc1eb79a7-kube-api-access-r2jbb\") pod \"010fee6b-bceb-4a25-b50f-19ccc1eb79a7\" (UID: \"010fee6b-bceb-4a25-b50f-19ccc1eb79a7\") " Oct 07 14:02:20 crc kubenswrapper[4959]: I1007 14:02:20.236402 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/010fee6b-bceb-4a25-b50f-19ccc1eb79a7-kube-api-access-r2jbb" (OuterVolumeSpecName: "kube-api-access-r2jbb") pod "010fee6b-bceb-4a25-b50f-19ccc1eb79a7" (UID: "010fee6b-bceb-4a25-b50f-19ccc1eb79a7"). InnerVolumeSpecName "kube-api-access-r2jbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:02:20 crc kubenswrapper[4959]: I1007 14:02:20.241962 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bea0796-b66c-4c44-be6e-840bde657893-kube-api-access-vd9bj" (OuterVolumeSpecName: "kube-api-access-vd9bj") pod "2bea0796-b66c-4c44-be6e-840bde657893" (UID: "2bea0796-b66c-4c44-be6e-840bde657893"). InnerVolumeSpecName "kube-api-access-vd9bj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:02:20 crc kubenswrapper[4959]: I1007 14:02:20.242023 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32df9b0d-1d2c-4724-b6c7-c4902baf8ae0-kube-api-access-48fcl" (OuterVolumeSpecName: "kube-api-access-48fcl") pod "32df9b0d-1d2c-4724-b6c7-c4902baf8ae0" (UID: "32df9b0d-1d2c-4724-b6c7-c4902baf8ae0"). InnerVolumeSpecName "kube-api-access-48fcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:02:20 crc kubenswrapper[4959]: I1007 14:02:20.332964 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48fcl\" (UniqueName: \"kubernetes.io/projected/32df9b0d-1d2c-4724-b6c7-c4902baf8ae0-kube-api-access-48fcl\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:20 crc kubenswrapper[4959]: I1007 14:02:20.332998 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vd9bj\" (UniqueName: \"kubernetes.io/projected/2bea0796-b66c-4c44-be6e-840bde657893-kube-api-access-vd9bj\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:20 crc kubenswrapper[4959]: I1007 14:02:20.333008 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2jbb\" (UniqueName: \"kubernetes.io/projected/010fee6b-bceb-4a25-b50f-19ccc1eb79a7-kube-api-access-r2jbb\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:20 crc kubenswrapper[4959]: I1007 14:02:20.537607 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-bpqp9" Oct 07 14:02:20 crc kubenswrapper[4959]: I1007 14:02:20.537631 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-bpqp9" event={"ID":"32df9b0d-1d2c-4724-b6c7-c4902baf8ae0","Type":"ContainerDied","Data":"9e35b933ecd6375c473514f409887a2aa1aff101cdf1c3d3988e6bd65c58d334"} Oct 07 14:02:20 crc kubenswrapper[4959]: I1007 14:02:20.537695 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e35b933ecd6375c473514f409887a2aa1aff101cdf1c3d3988e6bd65c58d334" Oct 07 14:02:20 crc kubenswrapper[4959]: I1007 14:02:20.540133 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qgn9g" event={"ID":"010fee6b-bceb-4a25-b50f-19ccc1eb79a7","Type":"ContainerDied","Data":"fb700398daebae30300b22b1e5c4abd82425bcf5c1ba67555c2dbade6520349e"} Oct 07 14:02:20 crc kubenswrapper[4959]: I1007 14:02:20.540195 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb700398daebae30300b22b1e5c4abd82425bcf5c1ba67555c2dbade6520349e" Oct 07 14:02:20 crc kubenswrapper[4959]: I1007 14:02:20.540146 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qgn9g" Oct 07 14:02:20 crc kubenswrapper[4959]: I1007 14:02:20.541570 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-btxq4" event={"ID":"2bea0796-b66c-4c44-be6e-840bde657893","Type":"ContainerDied","Data":"098d6ebdf347d3298668b83253b4529e7022d2be72c04ccd9d0d8661d4e2f002"} Oct 07 14:02:20 crc kubenswrapper[4959]: I1007 14:02:20.541602 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="098d6ebdf347d3298668b83253b4529e7022d2be72c04ccd9d0d8661d4e2f002" Oct 07 14:02:20 crc kubenswrapper[4959]: I1007 14:02:20.541652 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-btxq4" Oct 07 14:02:21 crc kubenswrapper[4959]: I1007 14:02:21.550842 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gvngb" event={"ID":"c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f","Type":"ContainerStarted","Data":"33bc5d4e608c9f7275ee45befb6853744451f5dffcc0951284c7709cfaced467"} Oct 07 14:02:21 crc kubenswrapper[4959]: I1007 14:02:21.567943 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-gvngb" podStartSLOduration=6.318357855 podStartE2EDuration="11.567923622s" podCreationTimestamp="2025-10-07 14:02:10 +0000 UTC" firstStartedPulling="2025-10-07 14:02:15.520639198 +0000 UTC m=+1017.604043523" lastFinishedPulling="2025-10-07 14:02:20.770204935 +0000 UTC m=+1022.853609290" observedRunningTime="2025-10-07 14:02:21.566465535 +0000 UTC m=+1023.649869860" watchObservedRunningTime="2025-10-07 14:02:21.567923622 +0000 UTC m=+1023.651327947" Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.567262 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-3409-account-create-fwppc"] Oct 07 14:02:30 crc kubenswrapper[4959]: E1007 14:02:30.568247 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32df9b0d-1d2c-4724-b6c7-c4902baf8ae0" containerName="mariadb-database-create" Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.568264 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="32df9b0d-1d2c-4724-b6c7-c4902baf8ae0" containerName="mariadb-database-create" Oct 07 14:02:30 crc kubenswrapper[4959]: E1007 14:02:30.568280 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bea0796-b66c-4c44-be6e-840bde657893" containerName="mariadb-database-create" Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.568285 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bea0796-b66c-4c44-be6e-840bde657893" containerName="mariadb-database-create" Oct 07 14:02:30 crc kubenswrapper[4959]: E1007 14:02:30.568297 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="010fee6b-bceb-4a25-b50f-19ccc1eb79a7" containerName="mariadb-database-create" Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.568304 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="010fee6b-bceb-4a25-b50f-19ccc1eb79a7" containerName="mariadb-database-create" Oct 07 14:02:30 crc kubenswrapper[4959]: E1007 14:02:30.568320 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f677e371-82ab-45eb-b358-a2a3ad838137" containerName="ovn-config" Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.568328 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f677e371-82ab-45eb-b358-a2a3ad838137" containerName="ovn-config" Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.568494 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f677e371-82ab-45eb-b358-a2a3ad838137" containerName="ovn-config" Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.568518 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bea0796-b66c-4c44-be6e-840bde657893" containerName="mariadb-database-create" Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.568536 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="32df9b0d-1d2c-4724-b6c7-c4902baf8ae0" containerName="mariadb-database-create" Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.568552 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="010fee6b-bceb-4a25-b50f-19ccc1eb79a7" containerName="mariadb-database-create" Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.569112 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3409-account-create-fwppc" Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.572341 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.578235 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-3409-account-create-fwppc"] Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.635452 4959 generic.go:334] "Generic (PLEG): container finished" podID="c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f" containerID="33bc5d4e608c9f7275ee45befb6853744451f5dffcc0951284c7709cfaced467" exitCode=0 Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.635505 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gvngb" event={"ID":"c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f","Type":"ContainerDied","Data":"33bc5d4e608c9f7275ee45befb6853744451f5dffcc0951284c7709cfaced467"} Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.670302 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-3d71-account-create-tpxcs"] Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.671430 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3d71-account-create-tpxcs" Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.677758 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.685731 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-3d71-account-create-tpxcs"] Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.715176 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc2sx\" (UniqueName: \"kubernetes.io/projected/56568d60-1394-4d21-af2c-2b5fa6bf8007-kube-api-access-fc2sx\") pod \"barbican-3409-account-create-fwppc\" (UID: \"56568d60-1394-4d21-af2c-2b5fa6bf8007\") " pod="openstack/barbican-3409-account-create-fwppc" Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.817001 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djcnf\" (UniqueName: \"kubernetes.io/projected/c88deaae-f6a8-45c8-b17b-6dc226c4f3d6-kube-api-access-djcnf\") pod \"cinder-3d71-account-create-tpxcs\" (UID: \"c88deaae-f6a8-45c8-b17b-6dc226c4f3d6\") " pod="openstack/cinder-3d71-account-create-tpxcs" Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.817086 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc2sx\" (UniqueName: \"kubernetes.io/projected/56568d60-1394-4d21-af2c-2b5fa6bf8007-kube-api-access-fc2sx\") pod \"barbican-3409-account-create-fwppc\" (UID: \"56568d60-1394-4d21-af2c-2b5fa6bf8007\") " pod="openstack/barbican-3409-account-create-fwppc" Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.845314 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc2sx\" (UniqueName: \"kubernetes.io/projected/56568d60-1394-4d21-af2c-2b5fa6bf8007-kube-api-access-fc2sx\") pod \"barbican-3409-account-create-fwppc\" (UID: \"56568d60-1394-4d21-af2c-2b5fa6bf8007\") " pod="openstack/barbican-3409-account-create-fwppc" Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.861973 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-d404-account-create-w2qcd"] Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.863566 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d404-account-create-w2qcd" Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.865835 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.880396 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d404-account-create-w2qcd"] Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.918001 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djcnf\" (UniqueName: \"kubernetes.io/projected/c88deaae-f6a8-45c8-b17b-6dc226c4f3d6-kube-api-access-djcnf\") pod \"cinder-3d71-account-create-tpxcs\" (UID: \"c88deaae-f6a8-45c8-b17b-6dc226c4f3d6\") " pod="openstack/cinder-3d71-account-create-tpxcs" Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.919545 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3409-account-create-fwppc" Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.934113 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djcnf\" (UniqueName: \"kubernetes.io/projected/c88deaae-f6a8-45c8-b17b-6dc226c4f3d6-kube-api-access-djcnf\") pod \"cinder-3d71-account-create-tpxcs\" (UID: \"c88deaae-f6a8-45c8-b17b-6dc226c4f3d6\") " pod="openstack/cinder-3d71-account-create-tpxcs" Oct 07 14:02:30 crc kubenswrapper[4959]: I1007 14:02:30.987128 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3d71-account-create-tpxcs" Oct 07 14:02:31 crc kubenswrapper[4959]: I1007 14:02:31.019403 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltmt8\" (UniqueName: \"kubernetes.io/projected/78ce8105-cd45-40fd-bfcf-2d853c8d822a-kube-api-access-ltmt8\") pod \"neutron-d404-account-create-w2qcd\" (UID: \"78ce8105-cd45-40fd-bfcf-2d853c8d822a\") " pod="openstack/neutron-d404-account-create-w2qcd" Oct 07 14:02:31 crc kubenswrapper[4959]: I1007 14:02:31.120980 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltmt8\" (UniqueName: \"kubernetes.io/projected/78ce8105-cd45-40fd-bfcf-2d853c8d822a-kube-api-access-ltmt8\") pod \"neutron-d404-account-create-w2qcd\" (UID: \"78ce8105-cd45-40fd-bfcf-2d853c8d822a\") " pod="openstack/neutron-d404-account-create-w2qcd" Oct 07 14:02:31 crc kubenswrapper[4959]: I1007 14:02:31.138484 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltmt8\" (UniqueName: \"kubernetes.io/projected/78ce8105-cd45-40fd-bfcf-2d853c8d822a-kube-api-access-ltmt8\") pod \"neutron-d404-account-create-w2qcd\" (UID: \"78ce8105-cd45-40fd-bfcf-2d853c8d822a\") " pod="openstack/neutron-d404-account-create-w2qcd" Oct 07 14:02:31 crc kubenswrapper[4959]: I1007 14:02:31.191594 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d404-account-create-w2qcd" Oct 07 14:02:31 crc kubenswrapper[4959]: I1007 14:02:31.404512 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d404-account-create-w2qcd"] Oct 07 14:02:31 crc kubenswrapper[4959]: W1007 14:02:31.404960 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78ce8105_cd45_40fd_bfcf_2d853c8d822a.slice/crio-8f217615923a593946e4dc5276fc7d4723546b24b0d5d8d2fb553ceccc962bb0 WatchSource:0}: Error finding container 8f217615923a593946e4dc5276fc7d4723546b24b0d5d8d2fb553ceccc962bb0: Status 404 returned error can't find the container with id 8f217615923a593946e4dc5276fc7d4723546b24b0d5d8d2fb553ceccc962bb0 Oct 07 14:02:31 crc kubenswrapper[4959]: I1007 14:02:31.420331 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-3409-account-create-fwppc"] Oct 07 14:02:31 crc kubenswrapper[4959]: W1007 14:02:31.432278 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56568d60_1394_4d21_af2c_2b5fa6bf8007.slice/crio-9485aa3ac5c6fdf036325b481581ae0a4a207bb6b30f8a7f21fff99dde35a8c7 WatchSource:0}: Error finding container 9485aa3ac5c6fdf036325b481581ae0a4a207bb6b30f8a7f21fff99dde35a8c7: Status 404 returned error can't find the container with id 9485aa3ac5c6fdf036325b481581ae0a4a207bb6b30f8a7f21fff99dde35a8c7 Oct 07 14:02:31 crc kubenswrapper[4959]: I1007 14:02:31.476795 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-3d71-account-create-tpxcs"] Oct 07 14:02:31 crc kubenswrapper[4959]: W1007 14:02:31.482838 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc88deaae_f6a8_45c8_b17b_6dc226c4f3d6.slice/crio-85d7c380286dff78f40b1bb5fb68da018f050c87007ffa57fdca02b0626d4bf4 WatchSource:0}: Error finding container 85d7c380286dff78f40b1bb5fb68da018f050c87007ffa57fdca02b0626d4bf4: Status 404 returned error can't find the container with id 85d7c380286dff78f40b1bb5fb68da018f050c87007ffa57fdca02b0626d4bf4 Oct 07 14:02:31 crc kubenswrapper[4959]: I1007 14:02:31.645771 4959 generic.go:334] "Generic (PLEG): container finished" podID="56568d60-1394-4d21-af2c-2b5fa6bf8007" containerID="a20ca8a566174703ad0d5a83c6ea9b5dd928950691431146eba300ba5a1de4ee" exitCode=0 Oct 07 14:02:31 crc kubenswrapper[4959]: I1007 14:02:31.645854 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3409-account-create-fwppc" event={"ID":"56568d60-1394-4d21-af2c-2b5fa6bf8007","Type":"ContainerDied","Data":"a20ca8a566174703ad0d5a83c6ea9b5dd928950691431146eba300ba5a1de4ee"} Oct 07 14:02:31 crc kubenswrapper[4959]: I1007 14:02:31.645908 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3409-account-create-fwppc" event={"ID":"56568d60-1394-4d21-af2c-2b5fa6bf8007","Type":"ContainerStarted","Data":"9485aa3ac5c6fdf036325b481581ae0a4a207bb6b30f8a7f21fff99dde35a8c7"} Oct 07 14:02:31 crc kubenswrapper[4959]: I1007 14:02:31.647786 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d404-account-create-w2qcd" event={"ID":"78ce8105-cd45-40fd-bfcf-2d853c8d822a","Type":"ContainerDied","Data":"4e442444929dab19f54564286b0dd42cc80b8aa37d28ea55436d6eba9233bdeb"} Oct 07 14:02:31 crc kubenswrapper[4959]: I1007 14:02:31.648387 4959 generic.go:334] "Generic (PLEG): container finished" podID="78ce8105-cd45-40fd-bfcf-2d853c8d822a" containerID="4e442444929dab19f54564286b0dd42cc80b8aa37d28ea55436d6eba9233bdeb" exitCode=0 Oct 07 14:02:31 crc kubenswrapper[4959]: I1007 14:02:31.648512 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d404-account-create-w2qcd" event={"ID":"78ce8105-cd45-40fd-bfcf-2d853c8d822a","Type":"ContainerStarted","Data":"8f217615923a593946e4dc5276fc7d4723546b24b0d5d8d2fb553ceccc962bb0"} Oct 07 14:02:31 crc kubenswrapper[4959]: I1007 14:02:31.653852 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3d71-account-create-tpxcs" event={"ID":"c88deaae-f6a8-45c8-b17b-6dc226c4f3d6","Type":"ContainerStarted","Data":"7e6ad7d9a1a80044b01306c65e9e6b1491b2359f709a14cc816497272c376c9b"} Oct 07 14:02:31 crc kubenswrapper[4959]: I1007 14:02:31.653878 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3d71-account-create-tpxcs" event={"ID":"c88deaae-f6a8-45c8-b17b-6dc226c4f3d6","Type":"ContainerStarted","Data":"85d7c380286dff78f40b1bb5fb68da018f050c87007ffa57fdca02b0626d4bf4"} Oct 07 14:02:31 crc kubenswrapper[4959]: I1007 14:02:31.688415 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-3d71-account-create-tpxcs" podStartSLOduration=1.688397838 podStartE2EDuration="1.688397838s" podCreationTimestamp="2025-10-07 14:02:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:02:31.682640433 +0000 UTC m=+1033.766044758" watchObservedRunningTime="2025-10-07 14:02:31.688397838 +0000 UTC m=+1033.771802163" Oct 07 14:02:31 crc kubenswrapper[4959]: I1007 14:02:31.984574 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gvngb" Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.040115 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxv2b\" (UniqueName: \"kubernetes.io/projected/c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f-kube-api-access-lxv2b\") pod \"c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f\" (UID: \"c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f\") " Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.040207 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f-combined-ca-bundle\") pod \"c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f\" (UID: \"c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f\") " Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.040273 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f-config-data\") pod \"c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f\" (UID: \"c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f\") " Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.047434 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f-kube-api-access-lxv2b" (OuterVolumeSpecName: "kube-api-access-lxv2b") pod "c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f" (UID: "c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f"). InnerVolumeSpecName "kube-api-access-lxv2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.063893 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f" (UID: "c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.097419 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f-config-data" (OuterVolumeSpecName: "config-data") pod "c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f" (UID: "c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.141742 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxv2b\" (UniqueName: \"kubernetes.io/projected/c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f-kube-api-access-lxv2b\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.141774 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.141785 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.667598 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gvngb" Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.670735 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gvngb" event={"ID":"c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f","Type":"ContainerDied","Data":"8cb665795ecb28d760aad58e90773b96ca5d9c24ce631440ad74f7bcdf5f16c7"} Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.670792 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8cb665795ecb28d760aad58e90773b96ca5d9c24ce631440ad74f7bcdf5f16c7" Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.675051 4959 generic.go:334] "Generic (PLEG): container finished" podID="024f57ee-b5d4-4f45-b447-5410cf1722ec" containerID="612295e06c3281aebb30eac6545902cd9ab7558de4fad84ee6b9eb0cbdc8d549" exitCode=0 Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.675145 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-fjrkz" event={"ID":"024f57ee-b5d4-4f45-b447-5410cf1722ec","Type":"ContainerDied","Data":"612295e06c3281aebb30eac6545902cd9ab7558de4fad84ee6b9eb0cbdc8d549"} Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.685027 4959 generic.go:334] "Generic (PLEG): container finished" podID="c88deaae-f6a8-45c8-b17b-6dc226c4f3d6" containerID="7e6ad7d9a1a80044b01306c65e9e6b1491b2359f709a14cc816497272c376c9b" exitCode=0 Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.685233 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3d71-account-create-tpxcs" event={"ID":"c88deaae-f6a8-45c8-b17b-6dc226c4f3d6","Type":"ContainerDied","Data":"7e6ad7d9a1a80044b01306c65e9e6b1491b2359f709a14cc816497272c376c9b"} Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.916076 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-v6lrw"] Oct 07 14:02:32 crc kubenswrapper[4959]: E1007 14:02:32.916645 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f" containerName="keystone-db-sync" Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.916660 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f" containerName="keystone-db-sync" Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.916835 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f" containerName="keystone-db-sync" Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.917329 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v6lrw" Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.922323 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5dcb7bb4dc-zz698"] Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.923542 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.923587 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dcb7bb4dc-zz698" Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.923744 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.923856 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.924063 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-z8j6b" Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.956903 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgjcq\" (UniqueName: \"kubernetes.io/projected/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-kube-api-access-fgjcq\") pod \"keystone-bootstrap-v6lrw\" (UID: \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\") " pod="openstack/keystone-bootstrap-v6lrw" Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.956946 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2qvh\" (UniqueName: \"kubernetes.io/projected/ee96a602-a1b4-4463-a919-b90fa83e23d3-kube-api-access-x2qvh\") pod \"dnsmasq-dns-5dcb7bb4dc-zz698\" (UID: \"ee96a602-a1b4-4463-a919-b90fa83e23d3\") " pod="openstack/dnsmasq-dns-5dcb7bb4dc-zz698" Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.956970 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee96a602-a1b4-4463-a919-b90fa83e23d3-dns-svc\") pod \"dnsmasq-dns-5dcb7bb4dc-zz698\" (UID: \"ee96a602-a1b4-4463-a919-b90fa83e23d3\") " pod="openstack/dnsmasq-dns-5dcb7bb4dc-zz698" Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.957008 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee96a602-a1b4-4463-a919-b90fa83e23d3-ovsdbserver-sb\") pod \"dnsmasq-dns-5dcb7bb4dc-zz698\" (UID: \"ee96a602-a1b4-4463-a919-b90fa83e23d3\") " pod="openstack/dnsmasq-dns-5dcb7bb4dc-zz698" Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.957071 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-config-data\") pod \"keystone-bootstrap-v6lrw\" (UID: \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\") " pod="openstack/keystone-bootstrap-v6lrw" Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.957091 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee96a602-a1b4-4463-a919-b90fa83e23d3-config\") pod \"dnsmasq-dns-5dcb7bb4dc-zz698\" (UID: \"ee96a602-a1b4-4463-a919-b90fa83e23d3\") " pod="openstack/dnsmasq-dns-5dcb7bb4dc-zz698" Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.957150 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-combined-ca-bundle\") pod \"keystone-bootstrap-v6lrw\" (UID: \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\") " pod="openstack/keystone-bootstrap-v6lrw" Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.957169 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee96a602-a1b4-4463-a919-b90fa83e23d3-ovsdbserver-nb\") pod \"dnsmasq-dns-5dcb7bb4dc-zz698\" (UID: \"ee96a602-a1b4-4463-a919-b90fa83e23d3\") " pod="openstack/dnsmasq-dns-5dcb7bb4dc-zz698" Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.957192 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-credential-keys\") pod \"keystone-bootstrap-v6lrw\" (UID: \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\") " pod="openstack/keystone-bootstrap-v6lrw" Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.957212 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-fernet-keys\") pod \"keystone-bootstrap-v6lrw\" (UID: \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\") " pod="openstack/keystone-bootstrap-v6lrw" Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.957235 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-scripts\") pod \"keystone-bootstrap-v6lrw\" (UID: \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\") " pod="openstack/keystone-bootstrap-v6lrw" Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.958684 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5dcb7bb4dc-zz698"] Oct 07 14:02:32 crc kubenswrapper[4959]: I1007 14:02:32.981729 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-v6lrw"] Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.060116 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-config-data\") pod \"keystone-bootstrap-v6lrw\" (UID: \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\") " pod="openstack/keystone-bootstrap-v6lrw" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.060163 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee96a602-a1b4-4463-a919-b90fa83e23d3-config\") pod \"dnsmasq-dns-5dcb7bb4dc-zz698\" (UID: \"ee96a602-a1b4-4463-a919-b90fa83e23d3\") " pod="openstack/dnsmasq-dns-5dcb7bb4dc-zz698" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.060208 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-combined-ca-bundle\") pod \"keystone-bootstrap-v6lrw\" (UID: \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\") " pod="openstack/keystone-bootstrap-v6lrw" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.060236 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee96a602-a1b4-4463-a919-b90fa83e23d3-ovsdbserver-nb\") pod \"dnsmasq-dns-5dcb7bb4dc-zz698\" (UID: \"ee96a602-a1b4-4463-a919-b90fa83e23d3\") " pod="openstack/dnsmasq-dns-5dcb7bb4dc-zz698" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.060263 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-credential-keys\") pod \"keystone-bootstrap-v6lrw\" (UID: \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\") " pod="openstack/keystone-bootstrap-v6lrw" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.060282 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-fernet-keys\") pod \"keystone-bootstrap-v6lrw\" (UID: \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\") " pod="openstack/keystone-bootstrap-v6lrw" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.060295 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-scripts\") pod \"keystone-bootstrap-v6lrw\" (UID: \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\") " pod="openstack/keystone-bootstrap-v6lrw" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.060317 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgjcq\" (UniqueName: \"kubernetes.io/projected/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-kube-api-access-fgjcq\") pod \"keystone-bootstrap-v6lrw\" (UID: \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\") " pod="openstack/keystone-bootstrap-v6lrw" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.060335 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2qvh\" (UniqueName: \"kubernetes.io/projected/ee96a602-a1b4-4463-a919-b90fa83e23d3-kube-api-access-x2qvh\") pod \"dnsmasq-dns-5dcb7bb4dc-zz698\" (UID: \"ee96a602-a1b4-4463-a919-b90fa83e23d3\") " pod="openstack/dnsmasq-dns-5dcb7bb4dc-zz698" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.060351 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee96a602-a1b4-4463-a919-b90fa83e23d3-dns-svc\") pod \"dnsmasq-dns-5dcb7bb4dc-zz698\" (UID: \"ee96a602-a1b4-4463-a919-b90fa83e23d3\") " pod="openstack/dnsmasq-dns-5dcb7bb4dc-zz698" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.060384 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee96a602-a1b4-4463-a919-b90fa83e23d3-ovsdbserver-sb\") pod \"dnsmasq-dns-5dcb7bb4dc-zz698\" (UID: \"ee96a602-a1b4-4463-a919-b90fa83e23d3\") " pod="openstack/dnsmasq-dns-5dcb7bb4dc-zz698" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.061368 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee96a602-a1b4-4463-a919-b90fa83e23d3-ovsdbserver-sb\") pod \"dnsmasq-dns-5dcb7bb4dc-zz698\" (UID: \"ee96a602-a1b4-4463-a919-b90fa83e23d3\") " pod="openstack/dnsmasq-dns-5dcb7bb4dc-zz698" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.066985 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee96a602-a1b4-4463-a919-b90fa83e23d3-config\") pod \"dnsmasq-dns-5dcb7bb4dc-zz698\" (UID: \"ee96a602-a1b4-4463-a919-b90fa83e23d3\") " pod="openstack/dnsmasq-dns-5dcb7bb4dc-zz698" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.070672 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-config-data\") pod \"keystone-bootstrap-v6lrw\" (UID: \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\") " pod="openstack/keystone-bootstrap-v6lrw" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.071275 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-scripts\") pod \"keystone-bootstrap-v6lrw\" (UID: \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\") " pod="openstack/keystone-bootstrap-v6lrw" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.072029 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-fernet-keys\") pod \"keystone-bootstrap-v6lrw\" (UID: \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\") " pod="openstack/keystone-bootstrap-v6lrw" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.086931 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee96a602-a1b4-4463-a919-b90fa83e23d3-dns-svc\") pod \"dnsmasq-dns-5dcb7bb4dc-zz698\" (UID: \"ee96a602-a1b4-4463-a919-b90fa83e23d3\") " pod="openstack/dnsmasq-dns-5dcb7bb4dc-zz698" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.087859 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee96a602-a1b4-4463-a919-b90fa83e23d3-ovsdbserver-nb\") pod \"dnsmasq-dns-5dcb7bb4dc-zz698\" (UID: \"ee96a602-a1b4-4463-a919-b90fa83e23d3\") " pod="openstack/dnsmasq-dns-5dcb7bb4dc-zz698" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.099630 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgjcq\" (UniqueName: \"kubernetes.io/projected/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-kube-api-access-fgjcq\") pod \"keystone-bootstrap-v6lrw\" (UID: \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\") " pod="openstack/keystone-bootstrap-v6lrw" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.102861 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2qvh\" (UniqueName: \"kubernetes.io/projected/ee96a602-a1b4-4463-a919-b90fa83e23d3-kube-api-access-x2qvh\") pod \"dnsmasq-dns-5dcb7bb4dc-zz698\" (UID: \"ee96a602-a1b4-4463-a919-b90fa83e23d3\") " pod="openstack/dnsmasq-dns-5dcb7bb4dc-zz698" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.103825 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-combined-ca-bundle\") pod \"keystone-bootstrap-v6lrw\" (UID: \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\") " pod="openstack/keystone-bootstrap-v6lrw" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.109839 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-credential-keys\") pod \"keystone-bootstrap-v6lrw\" (UID: \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\") " pod="openstack/keystone-bootstrap-v6lrw" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.171346 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-59bc478f75-xdfhn"] Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.172822 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-59bc478f75-xdfhn" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.175229 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-config-data\") pod \"horizon-59bc478f75-xdfhn\" (UID: \"5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0\") " pod="openstack/horizon-59bc478f75-xdfhn" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.175282 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-logs\") pod \"horizon-59bc478f75-xdfhn\" (UID: \"5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0\") " pod="openstack/horizon-59bc478f75-xdfhn" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.175352 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzcc6\" (UniqueName: \"kubernetes.io/projected/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-kube-api-access-jzcc6\") pod \"horizon-59bc478f75-xdfhn\" (UID: \"5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0\") " pod="openstack/horizon-59bc478f75-xdfhn" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.175421 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-scripts\") pod \"horizon-59bc478f75-xdfhn\" (UID: \"5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0\") " pod="openstack/horizon-59bc478f75-xdfhn" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.175440 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-horizon-secret-key\") pod \"horizon-59bc478f75-xdfhn\" (UID: \"5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0\") " pod="openstack/horizon-59bc478f75-xdfhn" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.179223 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.179558 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-cpdkq" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.180037 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.184770 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.196498 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-59bc478f75-xdfhn"] Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.241151 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v6lrw" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.242001 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-ff2nz"] Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.242945 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ff2nz" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.264391 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dcb7bb4dc-zz698" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.265064 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.265928 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.265944 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-lv2f8" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.279035 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-scripts\") pod \"horizon-59bc478f75-xdfhn\" (UID: \"5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0\") " pod="openstack/horizon-59bc478f75-xdfhn" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.279082 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-horizon-secret-key\") pod \"horizon-59bc478f75-xdfhn\" (UID: \"5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0\") " pod="openstack/horizon-59bc478f75-xdfhn" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.279124 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-config-data\") pod \"horizon-59bc478f75-xdfhn\" (UID: \"5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0\") " pod="openstack/horizon-59bc478f75-xdfhn" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.279151 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-logs\") pod \"horizon-59bc478f75-xdfhn\" (UID: \"5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0\") " pod="openstack/horizon-59bc478f75-xdfhn" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.279214 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzcc6\" (UniqueName: \"kubernetes.io/projected/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-kube-api-access-jzcc6\") pod \"horizon-59bc478f75-xdfhn\" (UID: \"5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0\") " pod="openstack/horizon-59bc478f75-xdfhn" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.279893 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-scripts\") pod \"horizon-59bc478f75-xdfhn\" (UID: \"5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0\") " pod="openstack/horizon-59bc478f75-xdfhn" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.281138 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-ff2nz"] Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.285584 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-horizon-secret-key\") pod \"horizon-59bc478f75-xdfhn\" (UID: \"5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0\") " pod="openstack/horizon-59bc478f75-xdfhn" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.291390 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-config-data\") pod \"horizon-59bc478f75-xdfhn\" (UID: \"5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0\") " pod="openstack/horizon-59bc478f75-xdfhn" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.301681 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-logs\") pod \"horizon-59bc478f75-xdfhn\" (UID: \"5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0\") " pod="openstack/horizon-59bc478f75-xdfhn" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.306370 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzcc6\" (UniqueName: \"kubernetes.io/projected/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-kube-api-access-jzcc6\") pod \"horizon-59bc478f75-xdfhn\" (UID: \"5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0\") " pod="openstack/horizon-59bc478f75-xdfhn" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.309505 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dcb7bb4dc-zz698"] Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.338240 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-d5f5b8fd9-mdlh9"] Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.339634 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d5f5b8fd9-mdlh9" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.367622 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3409-account-create-fwppc" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.367901 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-d5f5b8fd9-mdlh9"] Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.380310 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fcce932-e96f-4a33-9511-d84540bd98fa-combined-ca-bundle\") pod \"placement-db-sync-ff2nz\" (UID: \"2fcce932-e96f-4a33-9511-d84540bd98fa\") " pod="openstack/placement-db-sync-ff2nz" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.380359 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqwlb\" (UniqueName: \"kubernetes.io/projected/2fcce932-e96f-4a33-9511-d84540bd98fa-kube-api-access-dqwlb\") pod \"placement-db-sync-ff2nz\" (UID: \"2fcce932-e96f-4a33-9511-d84540bd98fa\") " pod="openstack/placement-db-sync-ff2nz" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.380397 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fcce932-e96f-4a33-9511-d84540bd98fa-logs\") pod \"placement-db-sync-ff2nz\" (UID: \"2fcce932-e96f-4a33-9511-d84540bd98fa\") " pod="openstack/placement-db-sync-ff2nz" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.380420 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fcce932-e96f-4a33-9511-d84540bd98fa-scripts\") pod \"placement-db-sync-ff2nz\" (UID: \"2fcce932-e96f-4a33-9511-d84540bd98fa\") " pod="openstack/placement-db-sync-ff2nz" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.380466 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fcce932-e96f-4a33-9511-d84540bd98fa-config-data\") pod \"placement-db-sync-ff2nz\" (UID: \"2fcce932-e96f-4a33-9511-d84540bd98fa\") " pod="openstack/placement-db-sync-ff2nz" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.384032 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f549f7b99-nhlx5"] Oct 07 14:02:33 crc kubenswrapper[4959]: E1007 14:02:33.384428 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56568d60-1394-4d21-af2c-2b5fa6bf8007" containerName="mariadb-account-create" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.384439 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="56568d60-1394-4d21-af2c-2b5fa6bf8007" containerName="mariadb-account-create" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.384597 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="56568d60-1394-4d21-af2c-2b5fa6bf8007" containerName="mariadb-account-create" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.385448 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f549f7b99-nhlx5" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.391774 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d404-account-create-w2qcd" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.416227 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f549f7b99-nhlx5"] Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.476230 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:02:33 crc kubenswrapper[4959]: E1007 14:02:33.477434 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78ce8105-cd45-40fd-bfcf-2d853c8d822a" containerName="mariadb-account-create" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.477459 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="78ce8105-cd45-40fd-bfcf-2d853c8d822a" containerName="mariadb-account-create" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.477880 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="78ce8105-cd45-40fd-bfcf-2d853c8d822a" containerName="mariadb-account-create" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.491200 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fc2sx\" (UniqueName: \"kubernetes.io/projected/56568d60-1394-4d21-af2c-2b5fa6bf8007-kube-api-access-fc2sx\") pod \"56568d60-1394-4d21-af2c-2b5fa6bf8007\" (UID: \"56568d60-1394-4d21-af2c-2b5fa6bf8007\") " Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.492622 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fcce932-e96f-4a33-9511-d84540bd98fa-config-data\") pod \"placement-db-sync-ff2nz\" (UID: \"2fcce932-e96f-4a33-9511-d84540bd98fa\") " pod="openstack/placement-db-sync-ff2nz" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.492670 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/906180a5-7c20-4109-8094-7d578d8ad374-config-data\") pod \"horizon-d5f5b8fd9-mdlh9\" (UID: \"906180a5-7c20-4109-8094-7d578d8ad374\") " pod="openstack/horizon-d5f5b8fd9-mdlh9" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.492706 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/553dac39-5e7d-4cd4-8f2d-e096814c34c7-config\") pod \"dnsmasq-dns-7f549f7b99-nhlx5\" (UID: \"553dac39-5e7d-4cd4-8f2d-e096814c34c7\") " pod="openstack/dnsmasq-dns-7f549f7b99-nhlx5" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.492742 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/906180a5-7c20-4109-8094-7d578d8ad374-scripts\") pod \"horizon-d5f5b8fd9-mdlh9\" (UID: \"906180a5-7c20-4109-8094-7d578d8ad374\") " pod="openstack/horizon-d5f5b8fd9-mdlh9" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.492799 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/906180a5-7c20-4109-8094-7d578d8ad374-logs\") pod \"horizon-d5f5b8fd9-mdlh9\" (UID: \"906180a5-7c20-4109-8094-7d578d8ad374\") " pod="openstack/horizon-d5f5b8fd9-mdlh9" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.492824 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/553dac39-5e7d-4cd4-8f2d-e096814c34c7-ovsdbserver-sb\") pod \"dnsmasq-dns-7f549f7b99-nhlx5\" (UID: \"553dac39-5e7d-4cd4-8f2d-e096814c34c7\") " pod="openstack/dnsmasq-dns-7f549f7b99-nhlx5" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.492861 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/553dac39-5e7d-4cd4-8f2d-e096814c34c7-dns-svc\") pod \"dnsmasq-dns-7f549f7b99-nhlx5\" (UID: \"553dac39-5e7d-4cd4-8f2d-e096814c34c7\") " pod="openstack/dnsmasq-dns-7f549f7b99-nhlx5" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.492910 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/906180a5-7c20-4109-8094-7d578d8ad374-horizon-secret-key\") pod \"horizon-d5f5b8fd9-mdlh9\" (UID: \"906180a5-7c20-4109-8094-7d578d8ad374\") " pod="openstack/horizon-d5f5b8fd9-mdlh9" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.492942 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fcce932-e96f-4a33-9511-d84540bd98fa-combined-ca-bundle\") pod \"placement-db-sync-ff2nz\" (UID: \"2fcce932-e96f-4a33-9511-d84540bd98fa\") " pod="openstack/placement-db-sync-ff2nz" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.493216 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqwlb\" (UniqueName: \"kubernetes.io/projected/2fcce932-e96f-4a33-9511-d84540bd98fa-kube-api-access-dqwlb\") pod \"placement-db-sync-ff2nz\" (UID: \"2fcce932-e96f-4a33-9511-d84540bd98fa\") " pod="openstack/placement-db-sync-ff2nz" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.493945 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.496678 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fcce932-e96f-4a33-9511-d84540bd98fa-logs\") pod \"placement-db-sync-ff2nz\" (UID: \"2fcce932-e96f-4a33-9511-d84540bd98fa\") " pod="openstack/placement-db-sync-ff2nz" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.496765 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fcce932-e96f-4a33-9511-d84540bd98fa-scripts\") pod \"placement-db-sync-ff2nz\" (UID: \"2fcce932-e96f-4a33-9511-d84540bd98fa\") " pod="openstack/placement-db-sync-ff2nz" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.502841 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56568d60-1394-4d21-af2c-2b5fa6bf8007-kube-api-access-fc2sx" (OuterVolumeSpecName: "kube-api-access-fc2sx") pod "56568d60-1394-4d21-af2c-2b5fa6bf8007" (UID: "56568d60-1394-4d21-af2c-2b5fa6bf8007"). InnerVolumeSpecName "kube-api-access-fc2sx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.511861 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fcce932-e96f-4a33-9511-d84540bd98fa-config-data\") pod \"placement-db-sync-ff2nz\" (UID: \"2fcce932-e96f-4a33-9511-d84540bd98fa\") " pod="openstack/placement-db-sync-ff2nz" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.512636 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fcce932-e96f-4a33-9511-d84540bd98fa-combined-ca-bundle\") pod \"placement-db-sync-ff2nz\" (UID: \"2fcce932-e96f-4a33-9511-d84540bd98fa\") " pod="openstack/placement-db-sync-ff2nz" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.512746 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.513803 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/553dac39-5e7d-4cd4-8f2d-e096814c34c7-ovsdbserver-nb\") pod \"dnsmasq-dns-7f549f7b99-nhlx5\" (UID: \"553dac39-5e7d-4cd4-8f2d-e096814c34c7\") " pod="openstack/dnsmasq-dns-7f549f7b99-nhlx5" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.513906 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6nvt\" (UniqueName: \"kubernetes.io/projected/906180a5-7c20-4109-8094-7d578d8ad374-kube-api-access-c6nvt\") pod \"horizon-d5f5b8fd9-mdlh9\" (UID: \"906180a5-7c20-4109-8094-7d578d8ad374\") " pod="openstack/horizon-d5f5b8fd9-mdlh9" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.513943 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8wxn\" (UniqueName: \"kubernetes.io/projected/553dac39-5e7d-4cd4-8f2d-e096814c34c7-kube-api-access-d8wxn\") pod \"dnsmasq-dns-7f549f7b99-nhlx5\" (UID: \"553dac39-5e7d-4cd4-8f2d-e096814c34c7\") " pod="openstack/dnsmasq-dns-7f549f7b99-nhlx5" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.529534 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fc2sx\" (UniqueName: \"kubernetes.io/projected/56568d60-1394-4d21-af2c-2b5fa6bf8007-kube-api-access-fc2sx\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.514472 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fcce932-e96f-4a33-9511-d84540bd98fa-logs\") pod \"placement-db-sync-ff2nz\" (UID: \"2fcce932-e96f-4a33-9511-d84540bd98fa\") " pod="openstack/placement-db-sync-ff2nz" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.535325 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqwlb\" (UniqueName: \"kubernetes.io/projected/2fcce932-e96f-4a33-9511-d84540bd98fa-kube-api-access-dqwlb\") pod \"placement-db-sync-ff2nz\" (UID: \"2fcce932-e96f-4a33-9511-d84540bd98fa\") " pod="openstack/placement-db-sync-ff2nz" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.535653 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.536356 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.539438 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fcce932-e96f-4a33-9511-d84540bd98fa-scripts\") pod \"placement-db-sync-ff2nz\" (UID: \"2fcce932-e96f-4a33-9511-d84540bd98fa\") " pod="openstack/placement-db-sync-ff2nz" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.564504 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-59bc478f75-xdfhn" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.630026 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltmt8\" (UniqueName: \"kubernetes.io/projected/78ce8105-cd45-40fd-bfcf-2d853c8d822a-kube-api-access-ltmt8\") pod \"78ce8105-cd45-40fd-bfcf-2d853c8d822a\" (UID: \"78ce8105-cd45-40fd-bfcf-2d853c8d822a\") " Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.630456 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/906180a5-7c20-4109-8094-7d578d8ad374-horizon-secret-key\") pod \"horizon-d5f5b8fd9-mdlh9\" (UID: \"906180a5-7c20-4109-8094-7d578d8ad374\") " pod="openstack/horizon-d5f5b8fd9-mdlh9" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.630537 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/553dac39-5e7d-4cd4-8f2d-e096814c34c7-ovsdbserver-nb\") pod \"dnsmasq-dns-7f549f7b99-nhlx5\" (UID: \"553dac39-5e7d-4cd4-8f2d-e096814c34c7\") " pod="openstack/dnsmasq-dns-7f549f7b99-nhlx5" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.630730 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6nvt\" (UniqueName: \"kubernetes.io/projected/906180a5-7c20-4109-8094-7d578d8ad374-kube-api-access-c6nvt\") pod \"horizon-d5f5b8fd9-mdlh9\" (UID: \"906180a5-7c20-4109-8094-7d578d8ad374\") " pod="openstack/horizon-d5f5b8fd9-mdlh9" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.630778 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8wxn\" (UniqueName: \"kubernetes.io/projected/553dac39-5e7d-4cd4-8f2d-e096814c34c7-kube-api-access-d8wxn\") pod \"dnsmasq-dns-7f549f7b99-nhlx5\" (UID: \"553dac39-5e7d-4cd4-8f2d-e096814c34c7\") " pod="openstack/dnsmasq-dns-7f549f7b99-nhlx5" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.630856 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/906180a5-7c20-4109-8094-7d578d8ad374-config-data\") pod \"horizon-d5f5b8fd9-mdlh9\" (UID: \"906180a5-7c20-4109-8094-7d578d8ad374\") " pod="openstack/horizon-d5f5b8fd9-mdlh9" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.630880 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/553dac39-5e7d-4cd4-8f2d-e096814c34c7-config\") pod \"dnsmasq-dns-7f549f7b99-nhlx5\" (UID: \"553dac39-5e7d-4cd4-8f2d-e096814c34c7\") " pod="openstack/dnsmasq-dns-7f549f7b99-nhlx5" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.630936 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/906180a5-7c20-4109-8094-7d578d8ad374-scripts\") pod \"horizon-d5f5b8fd9-mdlh9\" (UID: \"906180a5-7c20-4109-8094-7d578d8ad374\") " pod="openstack/horizon-d5f5b8fd9-mdlh9" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.631010 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/906180a5-7c20-4109-8094-7d578d8ad374-logs\") pod \"horizon-d5f5b8fd9-mdlh9\" (UID: \"906180a5-7c20-4109-8094-7d578d8ad374\") " pod="openstack/horizon-d5f5b8fd9-mdlh9" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.631030 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/553dac39-5e7d-4cd4-8f2d-e096814c34c7-ovsdbserver-sb\") pod \"dnsmasq-dns-7f549f7b99-nhlx5\" (UID: \"553dac39-5e7d-4cd4-8f2d-e096814c34c7\") " pod="openstack/dnsmasq-dns-7f549f7b99-nhlx5" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.631084 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/553dac39-5e7d-4cd4-8f2d-e096814c34c7-dns-svc\") pod \"dnsmasq-dns-7f549f7b99-nhlx5\" (UID: \"553dac39-5e7d-4cd4-8f2d-e096814c34c7\") " pod="openstack/dnsmasq-dns-7f549f7b99-nhlx5" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.631606 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/906180a5-7c20-4109-8094-7d578d8ad374-logs\") pod \"horizon-d5f5b8fd9-mdlh9\" (UID: \"906180a5-7c20-4109-8094-7d578d8ad374\") " pod="openstack/horizon-d5f5b8fd9-mdlh9" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.631611 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/906180a5-7c20-4109-8094-7d578d8ad374-scripts\") pod \"horizon-d5f5b8fd9-mdlh9\" (UID: \"906180a5-7c20-4109-8094-7d578d8ad374\") " pod="openstack/horizon-d5f5b8fd9-mdlh9" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.633306 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/906180a5-7c20-4109-8094-7d578d8ad374-config-data\") pod \"horizon-d5f5b8fd9-mdlh9\" (UID: \"906180a5-7c20-4109-8094-7d578d8ad374\") " pod="openstack/horizon-d5f5b8fd9-mdlh9" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.635375 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/553dac39-5e7d-4cd4-8f2d-e096814c34c7-ovsdbserver-nb\") pod \"dnsmasq-dns-7f549f7b99-nhlx5\" (UID: \"553dac39-5e7d-4cd4-8f2d-e096814c34c7\") " pod="openstack/dnsmasq-dns-7f549f7b99-nhlx5" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.635894 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/553dac39-5e7d-4cd4-8f2d-e096814c34c7-config\") pod \"dnsmasq-dns-7f549f7b99-nhlx5\" (UID: \"553dac39-5e7d-4cd4-8f2d-e096814c34c7\") " pod="openstack/dnsmasq-dns-7f549f7b99-nhlx5" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.635952 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/553dac39-5e7d-4cd4-8f2d-e096814c34c7-dns-svc\") pod \"dnsmasq-dns-7f549f7b99-nhlx5\" (UID: \"553dac39-5e7d-4cd4-8f2d-e096814c34c7\") " pod="openstack/dnsmasq-dns-7f549f7b99-nhlx5" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.637453 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/553dac39-5e7d-4cd4-8f2d-e096814c34c7-ovsdbserver-sb\") pod \"dnsmasq-dns-7f549f7b99-nhlx5\" (UID: \"553dac39-5e7d-4cd4-8f2d-e096814c34c7\") " pod="openstack/dnsmasq-dns-7f549f7b99-nhlx5" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.651329 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78ce8105-cd45-40fd-bfcf-2d853c8d822a-kube-api-access-ltmt8" (OuterVolumeSpecName: "kube-api-access-ltmt8") pod "78ce8105-cd45-40fd-bfcf-2d853c8d822a" (UID: "78ce8105-cd45-40fd-bfcf-2d853c8d822a"). InnerVolumeSpecName "kube-api-access-ltmt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.659706 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6nvt\" (UniqueName: \"kubernetes.io/projected/906180a5-7c20-4109-8094-7d578d8ad374-kube-api-access-c6nvt\") pod \"horizon-d5f5b8fd9-mdlh9\" (UID: \"906180a5-7c20-4109-8094-7d578d8ad374\") " pod="openstack/horizon-d5f5b8fd9-mdlh9" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.663238 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8wxn\" (UniqueName: \"kubernetes.io/projected/553dac39-5e7d-4cd4-8f2d-e096814c34c7-kube-api-access-d8wxn\") pod \"dnsmasq-dns-7f549f7b99-nhlx5\" (UID: \"553dac39-5e7d-4cd4-8f2d-e096814c34c7\") " pod="openstack/dnsmasq-dns-7f549f7b99-nhlx5" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.665529 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ff2nz" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.670636 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/906180a5-7c20-4109-8094-7d578d8ad374-horizon-secret-key\") pod \"horizon-d5f5b8fd9-mdlh9\" (UID: \"906180a5-7c20-4109-8094-7d578d8ad374\") " pod="openstack/horizon-d5f5b8fd9-mdlh9" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.709850 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3409-account-create-fwppc" event={"ID":"56568d60-1394-4d21-af2c-2b5fa6bf8007","Type":"ContainerDied","Data":"9485aa3ac5c6fdf036325b481581ae0a4a207bb6b30f8a7f21fff99dde35a8c7"} Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.710236 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9485aa3ac5c6fdf036325b481581ae0a4a207bb6b30f8a7f21fff99dde35a8c7" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.710324 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3409-account-create-fwppc" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.715818 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d404-account-create-w2qcd" event={"ID":"78ce8105-cd45-40fd-bfcf-2d853c8d822a","Type":"ContainerDied","Data":"8f217615923a593946e4dc5276fc7d4723546b24b0d5d8d2fb553ceccc962bb0"} Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.715863 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f217615923a593946e4dc5276fc7d4723546b24b0d5d8d2fb553ceccc962bb0" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.715975 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d404-account-create-w2qcd" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.718638 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d5f5b8fd9-mdlh9" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.730179 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f549f7b99-nhlx5" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.732501 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cd33b52c-3f7f-425e-aa42-7342343067a6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cd33b52c-3f7f-425e-aa42-7342343067a6\") " pod="openstack/ceilometer-0" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.732609 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd33b52c-3f7f-425e-aa42-7342343067a6-scripts\") pod \"ceilometer-0\" (UID: \"cd33b52c-3f7f-425e-aa42-7342343067a6\") " pod="openstack/ceilometer-0" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.732636 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcw9j\" (UniqueName: \"kubernetes.io/projected/cd33b52c-3f7f-425e-aa42-7342343067a6-kube-api-access-lcw9j\") pod \"ceilometer-0\" (UID: \"cd33b52c-3f7f-425e-aa42-7342343067a6\") " pod="openstack/ceilometer-0" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.732662 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd33b52c-3f7f-425e-aa42-7342343067a6-run-httpd\") pod \"ceilometer-0\" (UID: \"cd33b52c-3f7f-425e-aa42-7342343067a6\") " pod="openstack/ceilometer-0" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.732689 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd33b52c-3f7f-425e-aa42-7342343067a6-config-data\") pod \"ceilometer-0\" (UID: \"cd33b52c-3f7f-425e-aa42-7342343067a6\") " pod="openstack/ceilometer-0" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.732708 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd33b52c-3f7f-425e-aa42-7342343067a6-log-httpd\") pod \"ceilometer-0\" (UID: \"cd33b52c-3f7f-425e-aa42-7342343067a6\") " pod="openstack/ceilometer-0" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.732762 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd33b52c-3f7f-425e-aa42-7342343067a6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cd33b52c-3f7f-425e-aa42-7342343067a6\") " pod="openstack/ceilometer-0" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.732886 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltmt8\" (UniqueName: \"kubernetes.io/projected/78ce8105-cd45-40fd-bfcf-2d853c8d822a-kube-api-access-ltmt8\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.833531 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cd33b52c-3f7f-425e-aa42-7342343067a6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cd33b52c-3f7f-425e-aa42-7342343067a6\") " pod="openstack/ceilometer-0" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.833606 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd33b52c-3f7f-425e-aa42-7342343067a6-scripts\") pod \"ceilometer-0\" (UID: \"cd33b52c-3f7f-425e-aa42-7342343067a6\") " pod="openstack/ceilometer-0" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.833632 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcw9j\" (UniqueName: \"kubernetes.io/projected/cd33b52c-3f7f-425e-aa42-7342343067a6-kube-api-access-lcw9j\") pod \"ceilometer-0\" (UID: \"cd33b52c-3f7f-425e-aa42-7342343067a6\") " pod="openstack/ceilometer-0" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.833655 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd33b52c-3f7f-425e-aa42-7342343067a6-run-httpd\") pod \"ceilometer-0\" (UID: \"cd33b52c-3f7f-425e-aa42-7342343067a6\") " pod="openstack/ceilometer-0" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.833679 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd33b52c-3f7f-425e-aa42-7342343067a6-config-data\") pod \"ceilometer-0\" (UID: \"cd33b52c-3f7f-425e-aa42-7342343067a6\") " pod="openstack/ceilometer-0" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.833693 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd33b52c-3f7f-425e-aa42-7342343067a6-log-httpd\") pod \"ceilometer-0\" (UID: \"cd33b52c-3f7f-425e-aa42-7342343067a6\") " pod="openstack/ceilometer-0" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.833734 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd33b52c-3f7f-425e-aa42-7342343067a6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cd33b52c-3f7f-425e-aa42-7342343067a6\") " pod="openstack/ceilometer-0" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.836461 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd33b52c-3f7f-425e-aa42-7342343067a6-run-httpd\") pod \"ceilometer-0\" (UID: \"cd33b52c-3f7f-425e-aa42-7342343067a6\") " pod="openstack/ceilometer-0" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.836719 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd33b52c-3f7f-425e-aa42-7342343067a6-log-httpd\") pod \"ceilometer-0\" (UID: \"cd33b52c-3f7f-425e-aa42-7342343067a6\") " pod="openstack/ceilometer-0" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.841730 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd33b52c-3f7f-425e-aa42-7342343067a6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cd33b52c-3f7f-425e-aa42-7342343067a6\") " pod="openstack/ceilometer-0" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.847675 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd33b52c-3f7f-425e-aa42-7342343067a6-scripts\") pod \"ceilometer-0\" (UID: \"cd33b52c-3f7f-425e-aa42-7342343067a6\") " pod="openstack/ceilometer-0" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.848544 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd33b52c-3f7f-425e-aa42-7342343067a6-config-data\") pod \"ceilometer-0\" (UID: \"cd33b52c-3f7f-425e-aa42-7342343067a6\") " pod="openstack/ceilometer-0" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.855538 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcw9j\" (UniqueName: \"kubernetes.io/projected/cd33b52c-3f7f-425e-aa42-7342343067a6-kube-api-access-lcw9j\") pod \"ceilometer-0\" (UID: \"cd33b52c-3f7f-425e-aa42-7342343067a6\") " pod="openstack/ceilometer-0" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.887481 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cd33b52c-3f7f-425e-aa42-7342343067a6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cd33b52c-3f7f-425e-aa42-7342343067a6\") " pod="openstack/ceilometer-0" Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.946005 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dcb7bb4dc-zz698"] Oct 07 14:02:33 crc kubenswrapper[4959]: W1007 14:02:33.956446 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee96a602_a1b4_4463_a919_b90fa83e23d3.slice/crio-4dfb81a288109f1aa6ac4cfb60e430ffe49364b0c3d4de9a8c7f7c671b3df005 WatchSource:0}: Error finding container 4dfb81a288109f1aa6ac4cfb60e430ffe49364b0c3d4de9a8c7f7c671b3df005: Status 404 returned error can't find the container with id 4dfb81a288109f1aa6ac4cfb60e430ffe49364b0c3d4de9a8c7f7c671b3df005 Oct 07 14:02:33 crc kubenswrapper[4959]: I1007 14:02:33.977399 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-v6lrw"] Oct 07 14:02:33 crc kubenswrapper[4959]: W1007 14:02:33.988066 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3205491a_c78b_4eff_b2e1_aa17eb81f5b4.slice/crio-f635c32fa93ab50f3a6dcf3a67212f60c73f8dadab7fd02cd593a5d8f254d24b WatchSource:0}: Error finding container f635c32fa93ab50f3a6dcf3a67212f60c73f8dadab7fd02cd593a5d8f254d24b: Status 404 returned error can't find the container with id f635c32fa93ab50f3a6dcf3a67212f60c73f8dadab7fd02cd593a5d8f254d24b Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.157538 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.284293 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3d71-account-create-tpxcs" Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.343538 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-ff2nz"] Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.365608 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-59bc478f75-xdfhn"] Oct 07 14:02:34 crc kubenswrapper[4959]: W1007 14:02:34.368092 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2fcce932_e96f_4a33_9511_d84540bd98fa.slice/crio-c5099f9b7273b37cc0577cb3ab3f9a6765d4643ebfca51d912543792e8040529 WatchSource:0}: Error finding container c5099f9b7273b37cc0577cb3ab3f9a6765d4643ebfca51d912543792e8040529: Status 404 returned error can't find the container with id c5099f9b7273b37cc0577cb3ab3f9a6765d4643ebfca51d912543792e8040529 Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.394272 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-fjrkz" Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.445760 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djcnf\" (UniqueName: \"kubernetes.io/projected/c88deaae-f6a8-45c8-b17b-6dc226c4f3d6-kube-api-access-djcnf\") pod \"c88deaae-f6a8-45c8-b17b-6dc226c4f3d6\" (UID: \"c88deaae-f6a8-45c8-b17b-6dc226c4f3d6\") " Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.452908 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c88deaae-f6a8-45c8-b17b-6dc226c4f3d6-kube-api-access-djcnf" (OuterVolumeSpecName: "kube-api-access-djcnf") pod "c88deaae-f6a8-45c8-b17b-6dc226c4f3d6" (UID: "c88deaae-f6a8-45c8-b17b-6dc226c4f3d6"). InnerVolumeSpecName "kube-api-access-djcnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.458837 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-d5f5b8fd9-mdlh9"] Oct 07 14:02:34 crc kubenswrapper[4959]: W1007 14:02:34.466286 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod906180a5_7c20_4109_8094_7d578d8ad374.slice/crio-21cddbd5537b4e7e16caebd3f512d0d4aac613cb8cc0b2a4abd593bb11f8e3a2 WatchSource:0}: Error finding container 21cddbd5537b4e7e16caebd3f512d0d4aac613cb8cc0b2a4abd593bb11f8e3a2: Status 404 returned error can't find the container with id 21cddbd5537b4e7e16caebd3f512d0d4aac613cb8cc0b2a4abd593bb11f8e3a2 Oct 07 14:02:34 crc kubenswrapper[4959]: W1007 14:02:34.479827 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod553dac39_5e7d_4cd4_8f2d_e096814c34c7.slice/crio-2ea96f487963ecf53d1963a52fcce8fd3233c7f69f727aa6085a7bcfc276a13d WatchSource:0}: Error finding container 2ea96f487963ecf53d1963a52fcce8fd3233c7f69f727aa6085a7bcfc276a13d: Status 404 returned error can't find the container with id 2ea96f487963ecf53d1963a52fcce8fd3233c7f69f727aa6085a7bcfc276a13d Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.487375 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f549f7b99-nhlx5"] Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.548357 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/024f57ee-b5d4-4f45-b447-5410cf1722ec-config-data\") pod \"024f57ee-b5d4-4f45-b447-5410cf1722ec\" (UID: \"024f57ee-b5d4-4f45-b447-5410cf1722ec\") " Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.548447 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/024f57ee-b5d4-4f45-b447-5410cf1722ec-db-sync-config-data\") pod \"024f57ee-b5d4-4f45-b447-5410cf1722ec\" (UID: \"024f57ee-b5d4-4f45-b447-5410cf1722ec\") " Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.548579 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/024f57ee-b5d4-4f45-b447-5410cf1722ec-combined-ca-bundle\") pod \"024f57ee-b5d4-4f45-b447-5410cf1722ec\" (UID: \"024f57ee-b5d4-4f45-b447-5410cf1722ec\") " Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.548625 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7sq2\" (UniqueName: \"kubernetes.io/projected/024f57ee-b5d4-4f45-b447-5410cf1722ec-kube-api-access-k7sq2\") pod \"024f57ee-b5d4-4f45-b447-5410cf1722ec\" (UID: \"024f57ee-b5d4-4f45-b447-5410cf1722ec\") " Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.548907 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djcnf\" (UniqueName: \"kubernetes.io/projected/c88deaae-f6a8-45c8-b17b-6dc226c4f3d6-kube-api-access-djcnf\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.552255 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/024f57ee-b5d4-4f45-b447-5410cf1722ec-kube-api-access-k7sq2" (OuterVolumeSpecName: "kube-api-access-k7sq2") pod "024f57ee-b5d4-4f45-b447-5410cf1722ec" (UID: "024f57ee-b5d4-4f45-b447-5410cf1722ec"). InnerVolumeSpecName "kube-api-access-k7sq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.552267 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/024f57ee-b5d4-4f45-b447-5410cf1722ec-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "024f57ee-b5d4-4f45-b447-5410cf1722ec" (UID: "024f57ee-b5d4-4f45-b447-5410cf1722ec"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.572818 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/024f57ee-b5d4-4f45-b447-5410cf1722ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "024f57ee-b5d4-4f45-b447-5410cf1722ec" (UID: "024f57ee-b5d4-4f45-b447-5410cf1722ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.599409 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/024f57ee-b5d4-4f45-b447-5410cf1722ec-config-data" (OuterVolumeSpecName: "config-data") pod "024f57ee-b5d4-4f45-b447-5410cf1722ec" (UID: "024f57ee-b5d4-4f45-b447-5410cf1722ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.651828 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/024f57ee-b5d4-4f45-b447-5410cf1722ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.651867 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7sq2\" (UniqueName: \"kubernetes.io/projected/024f57ee-b5d4-4f45-b447-5410cf1722ec-kube-api-access-k7sq2\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.651884 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/024f57ee-b5d4-4f45-b447-5410cf1722ec-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.651899 4959 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/024f57ee-b5d4-4f45-b447-5410cf1722ec-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:34 crc kubenswrapper[4959]: W1007 14:02:34.687457 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd33b52c_3f7f_425e_aa42_7342343067a6.slice/crio-4a10abf5a6e5c04f37c12657f3cc13c27886b75c4c9951e8a873277487ae2890 WatchSource:0}: Error finding container 4a10abf5a6e5c04f37c12657f3cc13c27886b75c4c9951e8a873277487ae2890: Status 404 returned error can't find the container with id 4a10abf5a6e5c04f37c12657f3cc13c27886b75c4c9951e8a873277487ae2890 Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.696184 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.738132 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v6lrw" event={"ID":"3205491a-c78b-4eff-b2e1-aa17eb81f5b4","Type":"ContainerStarted","Data":"f635c32fa93ab50f3a6dcf3a67212f60c73f8dadab7fd02cd593a5d8f254d24b"} Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.742643 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3d71-account-create-tpxcs" Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.742666 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3d71-account-create-tpxcs" event={"ID":"c88deaae-f6a8-45c8-b17b-6dc226c4f3d6","Type":"ContainerDied","Data":"85d7c380286dff78f40b1bb5fb68da018f050c87007ffa57fdca02b0626d4bf4"} Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.742736 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85d7c380286dff78f40b1bb5fb68da018f050c87007ffa57fdca02b0626d4bf4" Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.747811 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ff2nz" event={"ID":"2fcce932-e96f-4a33-9511-d84540bd98fa","Type":"ContainerStarted","Data":"c5099f9b7273b37cc0577cb3ab3f9a6765d4643ebfca51d912543792e8040529"} Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.749940 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dcb7bb4dc-zz698" event={"ID":"ee96a602-a1b4-4463-a919-b90fa83e23d3","Type":"ContainerStarted","Data":"4dfb81a288109f1aa6ac4cfb60e430ffe49364b0c3d4de9a8c7f7c671b3df005"} Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.751566 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d5f5b8fd9-mdlh9" event={"ID":"906180a5-7c20-4109-8094-7d578d8ad374","Type":"ContainerStarted","Data":"21cddbd5537b4e7e16caebd3f512d0d4aac613cb8cc0b2a4abd593bb11f8e3a2"} Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.757338 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-59bc478f75-xdfhn" event={"ID":"5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0","Type":"ContainerStarted","Data":"3e92680d8b5a23c3bae9589d5d17b34b82712d00baf737b9f3d1445ec475e0c0"} Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.763008 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-fjrkz" event={"ID":"024f57ee-b5d4-4f45-b447-5410cf1722ec","Type":"ContainerDied","Data":"09880e9a444d85aca439b39c442908a1c3f79e08dfa55f047365308af4869ab2"} Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.763046 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09880e9a444d85aca439b39c442908a1c3f79e08dfa55f047365308af4869ab2" Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.763130 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-fjrkz" Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.767002 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd33b52c-3f7f-425e-aa42-7342343067a6","Type":"ContainerStarted","Data":"4a10abf5a6e5c04f37c12657f3cc13c27886b75c4c9951e8a873277487ae2890"} Oct 07 14:02:34 crc kubenswrapper[4959]: I1007 14:02:34.768044 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f549f7b99-nhlx5" event={"ID":"553dac39-5e7d-4cd4-8f2d-e096814c34c7","Type":"ContainerStarted","Data":"2ea96f487963ecf53d1963a52fcce8fd3233c7f69f727aa6085a7bcfc276a13d"} Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.036738 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f549f7b99-nhlx5"] Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.058534 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-748d7644cf-6zpv4"] Oct 07 14:02:35 crc kubenswrapper[4959]: E1007 14:02:35.059679 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="024f57ee-b5d4-4f45-b447-5410cf1722ec" containerName="glance-db-sync" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.059699 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="024f57ee-b5d4-4f45-b447-5410cf1722ec" containerName="glance-db-sync" Oct 07 14:02:35 crc kubenswrapper[4959]: E1007 14:02:35.059712 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c88deaae-f6a8-45c8-b17b-6dc226c4f3d6" containerName="mariadb-account-create" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.059722 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c88deaae-f6a8-45c8-b17b-6dc226c4f3d6" containerName="mariadb-account-create" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.059927 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="024f57ee-b5d4-4f45-b447-5410cf1722ec" containerName="glance-db-sync" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.059945 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="c88deaae-f6a8-45c8-b17b-6dc226c4f3d6" containerName="mariadb-account-create" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.060907 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-748d7644cf-6zpv4" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.097907 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-748d7644cf-6zpv4"] Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.168232 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ff263db-2611-4310-a9ee-01e2136c8b1b-dns-svc\") pod \"dnsmasq-dns-748d7644cf-6zpv4\" (UID: \"0ff263db-2611-4310-a9ee-01e2136c8b1b\") " pod="openstack/dnsmasq-dns-748d7644cf-6zpv4" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.168282 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ff263db-2611-4310-a9ee-01e2136c8b1b-ovsdbserver-nb\") pod \"dnsmasq-dns-748d7644cf-6zpv4\" (UID: \"0ff263db-2611-4310-a9ee-01e2136c8b1b\") " pod="openstack/dnsmasq-dns-748d7644cf-6zpv4" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.168332 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ff263db-2611-4310-a9ee-01e2136c8b1b-ovsdbserver-sb\") pod \"dnsmasq-dns-748d7644cf-6zpv4\" (UID: \"0ff263db-2611-4310-a9ee-01e2136c8b1b\") " pod="openstack/dnsmasq-dns-748d7644cf-6zpv4" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.168453 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m25xd\" (UniqueName: \"kubernetes.io/projected/0ff263db-2611-4310-a9ee-01e2136c8b1b-kube-api-access-m25xd\") pod \"dnsmasq-dns-748d7644cf-6zpv4\" (UID: \"0ff263db-2611-4310-a9ee-01e2136c8b1b\") " pod="openstack/dnsmasq-dns-748d7644cf-6zpv4" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.168542 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ff263db-2611-4310-a9ee-01e2136c8b1b-config\") pod \"dnsmasq-dns-748d7644cf-6zpv4\" (UID: \"0ff263db-2611-4310-a9ee-01e2136c8b1b\") " pod="openstack/dnsmasq-dns-748d7644cf-6zpv4" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.228367 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-59bc478f75-xdfhn"] Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.268978 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.269890 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m25xd\" (UniqueName: \"kubernetes.io/projected/0ff263db-2611-4310-a9ee-01e2136c8b1b-kube-api-access-m25xd\") pod \"dnsmasq-dns-748d7644cf-6zpv4\" (UID: \"0ff263db-2611-4310-a9ee-01e2136c8b1b\") " pod="openstack/dnsmasq-dns-748d7644cf-6zpv4" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.269963 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ff263db-2611-4310-a9ee-01e2136c8b1b-config\") pod \"dnsmasq-dns-748d7644cf-6zpv4\" (UID: \"0ff263db-2611-4310-a9ee-01e2136c8b1b\") " pod="openstack/dnsmasq-dns-748d7644cf-6zpv4" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.270019 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ff263db-2611-4310-a9ee-01e2136c8b1b-dns-svc\") pod \"dnsmasq-dns-748d7644cf-6zpv4\" (UID: \"0ff263db-2611-4310-a9ee-01e2136c8b1b\") " pod="openstack/dnsmasq-dns-748d7644cf-6zpv4" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.270049 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ff263db-2611-4310-a9ee-01e2136c8b1b-ovsdbserver-nb\") pod \"dnsmasq-dns-748d7644cf-6zpv4\" (UID: \"0ff263db-2611-4310-a9ee-01e2136c8b1b\") " pod="openstack/dnsmasq-dns-748d7644cf-6zpv4" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.270082 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ff263db-2611-4310-a9ee-01e2136c8b1b-ovsdbserver-sb\") pod \"dnsmasq-dns-748d7644cf-6zpv4\" (UID: \"0ff263db-2611-4310-a9ee-01e2136c8b1b\") " pod="openstack/dnsmasq-dns-748d7644cf-6zpv4" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.271262 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ff263db-2611-4310-a9ee-01e2136c8b1b-ovsdbserver-sb\") pod \"dnsmasq-dns-748d7644cf-6zpv4\" (UID: \"0ff263db-2611-4310-a9ee-01e2136c8b1b\") " pod="openstack/dnsmasq-dns-748d7644cf-6zpv4" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.271841 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ff263db-2611-4310-a9ee-01e2136c8b1b-ovsdbserver-nb\") pod \"dnsmasq-dns-748d7644cf-6zpv4\" (UID: \"0ff263db-2611-4310-a9ee-01e2136c8b1b\") " pod="openstack/dnsmasq-dns-748d7644cf-6zpv4" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.272814 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ff263db-2611-4310-a9ee-01e2136c8b1b-config\") pod \"dnsmasq-dns-748d7644cf-6zpv4\" (UID: \"0ff263db-2611-4310-a9ee-01e2136c8b1b\") " pod="openstack/dnsmasq-dns-748d7644cf-6zpv4" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.273373 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ff263db-2611-4310-a9ee-01e2136c8b1b-dns-svc\") pod \"dnsmasq-dns-748d7644cf-6zpv4\" (UID: \"0ff263db-2611-4310-a9ee-01e2136c8b1b\") " pod="openstack/dnsmasq-dns-748d7644cf-6zpv4" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.281002 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5f459969c7-55hf7"] Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.282290 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f459969c7-55hf7" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.286655 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5f459969c7-55hf7"] Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.295987 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m25xd\" (UniqueName: \"kubernetes.io/projected/0ff263db-2611-4310-a9ee-01e2136c8b1b-kube-api-access-m25xd\") pod \"dnsmasq-dns-748d7644cf-6zpv4\" (UID: \"0ff263db-2611-4310-a9ee-01e2136c8b1b\") " pod="openstack/dnsmasq-dns-748d7644cf-6zpv4" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.457290 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-748d7644cf-6zpv4" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.472463 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/594da4a6-8781-416f-abdc-8f694948b6a2-horizon-secret-key\") pod \"horizon-5f459969c7-55hf7\" (UID: \"594da4a6-8781-416f-abdc-8f694948b6a2\") " pod="openstack/horizon-5f459969c7-55hf7" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.472756 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/594da4a6-8781-416f-abdc-8f694948b6a2-scripts\") pod \"horizon-5f459969c7-55hf7\" (UID: \"594da4a6-8781-416f-abdc-8f694948b6a2\") " pod="openstack/horizon-5f459969c7-55hf7" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.472791 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/594da4a6-8781-416f-abdc-8f694948b6a2-logs\") pod \"horizon-5f459969c7-55hf7\" (UID: \"594da4a6-8781-416f-abdc-8f694948b6a2\") " pod="openstack/horizon-5f459969c7-55hf7" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.472837 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/594da4a6-8781-416f-abdc-8f694948b6a2-config-data\") pod \"horizon-5f459969c7-55hf7\" (UID: \"594da4a6-8781-416f-abdc-8f694948b6a2\") " pod="openstack/horizon-5f459969c7-55hf7" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.472960 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7x4vd\" (UniqueName: \"kubernetes.io/projected/594da4a6-8781-416f-abdc-8f694948b6a2-kube-api-access-7x4vd\") pod \"horizon-5f459969c7-55hf7\" (UID: \"594da4a6-8781-416f-abdc-8f694948b6a2\") " pod="openstack/horizon-5f459969c7-55hf7" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.573987 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/594da4a6-8781-416f-abdc-8f694948b6a2-config-data\") pod \"horizon-5f459969c7-55hf7\" (UID: \"594da4a6-8781-416f-abdc-8f694948b6a2\") " pod="openstack/horizon-5f459969c7-55hf7" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.574135 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7x4vd\" (UniqueName: \"kubernetes.io/projected/594da4a6-8781-416f-abdc-8f694948b6a2-kube-api-access-7x4vd\") pod \"horizon-5f459969c7-55hf7\" (UID: \"594da4a6-8781-416f-abdc-8f694948b6a2\") " pod="openstack/horizon-5f459969c7-55hf7" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.574166 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/594da4a6-8781-416f-abdc-8f694948b6a2-horizon-secret-key\") pod \"horizon-5f459969c7-55hf7\" (UID: \"594da4a6-8781-416f-abdc-8f694948b6a2\") " pod="openstack/horizon-5f459969c7-55hf7" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.574215 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/594da4a6-8781-416f-abdc-8f694948b6a2-scripts\") pod \"horizon-5f459969c7-55hf7\" (UID: \"594da4a6-8781-416f-abdc-8f694948b6a2\") " pod="openstack/horizon-5f459969c7-55hf7" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.574237 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/594da4a6-8781-416f-abdc-8f694948b6a2-logs\") pod \"horizon-5f459969c7-55hf7\" (UID: \"594da4a6-8781-416f-abdc-8f694948b6a2\") " pod="openstack/horizon-5f459969c7-55hf7" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.574748 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/594da4a6-8781-416f-abdc-8f694948b6a2-logs\") pod \"horizon-5f459969c7-55hf7\" (UID: \"594da4a6-8781-416f-abdc-8f694948b6a2\") " pod="openstack/horizon-5f459969c7-55hf7" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.574972 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/594da4a6-8781-416f-abdc-8f694948b6a2-scripts\") pod \"horizon-5f459969c7-55hf7\" (UID: \"594da4a6-8781-416f-abdc-8f694948b6a2\") " pod="openstack/horizon-5f459969c7-55hf7" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.581047 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/594da4a6-8781-416f-abdc-8f694948b6a2-horizon-secret-key\") pod \"horizon-5f459969c7-55hf7\" (UID: \"594da4a6-8781-416f-abdc-8f694948b6a2\") " pod="openstack/horizon-5f459969c7-55hf7" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.583134 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/594da4a6-8781-416f-abdc-8f694948b6a2-config-data\") pod \"horizon-5f459969c7-55hf7\" (UID: \"594da4a6-8781-416f-abdc-8f694948b6a2\") " pod="openstack/horizon-5f459969c7-55hf7" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.606664 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7x4vd\" (UniqueName: \"kubernetes.io/projected/594da4a6-8781-416f-abdc-8f694948b6a2-kube-api-access-7x4vd\") pod \"horizon-5f459969c7-55hf7\" (UID: \"594da4a6-8781-416f-abdc-8f694948b6a2\") " pod="openstack/horizon-5f459969c7-55hf7" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.634915 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f459969c7-55hf7" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.907499 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-748d7644cf-6zpv4"] Oct 07 14:02:35 crc kubenswrapper[4959]: W1007 14:02:35.920415 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ff263db_2611_4310_a9ee_01e2136c8b1b.slice/crio-adc28831f6472e2be951a70170e0355baa89eb96c0882c12dc2684466616c423 WatchSource:0}: Error finding container adc28831f6472e2be951a70170e0355baa89eb96c0882c12dc2684466616c423: Status 404 returned error can't find the container with id adc28831f6472e2be951a70170e0355baa89eb96c0882c12dc2684466616c423 Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.934163 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-2hhl9"] Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.935590 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2hhl9" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.940639 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.940837 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-tgdwf" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.950383 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-2hhl9"] Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.974567 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-jfrdg"] Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.976927 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-jfrdg" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.979908 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-k5lq9" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.979979 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.980016 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 07 14:02:35 crc kubenswrapper[4959]: I1007 14:02:35.987048 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-jfrdg"] Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.083969 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5f459969c7-55hf7"] Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.084187 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c3b8b961-3c0e-47dc-b28f-d55c9821ab34-db-sync-config-data\") pod \"barbican-db-sync-2hhl9\" (UID: \"c3b8b961-3c0e-47dc-b28f-d55c9821ab34\") " pod="openstack/barbican-db-sync-2hhl9" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.084241 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b70a321-ffb7-429f-b825-4cd872ded9ff-scripts\") pod \"cinder-db-sync-jfrdg\" (UID: \"7b70a321-ffb7-429f-b825-4cd872ded9ff\") " pod="openstack/cinder-db-sync-jfrdg" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.084270 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b70a321-ffb7-429f-b825-4cd872ded9ff-config-data\") pod \"cinder-db-sync-jfrdg\" (UID: \"7b70a321-ffb7-429f-b825-4cd872ded9ff\") " pod="openstack/cinder-db-sync-jfrdg" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.084308 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b70a321-ffb7-429f-b825-4cd872ded9ff-combined-ca-bundle\") pod \"cinder-db-sync-jfrdg\" (UID: \"7b70a321-ffb7-429f-b825-4cd872ded9ff\") " pod="openstack/cinder-db-sync-jfrdg" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.084332 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7b70a321-ffb7-429f-b825-4cd872ded9ff-db-sync-config-data\") pod \"cinder-db-sync-jfrdg\" (UID: \"7b70a321-ffb7-429f-b825-4cd872ded9ff\") " pod="openstack/cinder-db-sync-jfrdg" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.084366 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hw67n\" (UniqueName: \"kubernetes.io/projected/c3b8b961-3c0e-47dc-b28f-d55c9821ab34-kube-api-access-hw67n\") pod \"barbican-db-sync-2hhl9\" (UID: \"c3b8b961-3c0e-47dc-b28f-d55c9821ab34\") " pod="openstack/barbican-db-sync-2hhl9" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.084433 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3b8b961-3c0e-47dc-b28f-d55c9821ab34-combined-ca-bundle\") pod \"barbican-db-sync-2hhl9\" (UID: \"c3b8b961-3c0e-47dc-b28f-d55c9821ab34\") " pod="openstack/barbican-db-sync-2hhl9" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.084511 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmzf9\" (UniqueName: \"kubernetes.io/projected/7b70a321-ffb7-429f-b825-4cd872ded9ff-kube-api-access-zmzf9\") pod \"cinder-db-sync-jfrdg\" (UID: \"7b70a321-ffb7-429f-b825-4cd872ded9ff\") " pod="openstack/cinder-db-sync-jfrdg" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.084553 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7b70a321-ffb7-429f-b825-4cd872ded9ff-etc-machine-id\") pod \"cinder-db-sync-jfrdg\" (UID: \"7b70a321-ffb7-429f-b825-4cd872ded9ff\") " pod="openstack/cinder-db-sync-jfrdg" Oct 07 14:02:36 crc kubenswrapper[4959]: W1007 14:02:36.085632 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod594da4a6_8781_416f_abdc_8f694948b6a2.slice/crio-ef90acde1c02f78da6ee25dd4a23b978ff700b3a89c4af4bd1a7781af60004f0 WatchSource:0}: Error finding container ef90acde1c02f78da6ee25dd4a23b978ff700b3a89c4af4bd1a7781af60004f0: Status 404 returned error can't find the container with id ef90acde1c02f78da6ee25dd4a23b978ff700b3a89c4af4bd1a7781af60004f0 Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.172276 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-4sfdr"] Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.175374 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4sfdr" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.181070 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-rbsgf" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.181087 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.181087 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.183183 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-4sfdr"] Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.185794 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c3b8b961-3c0e-47dc-b28f-d55c9821ab34-db-sync-config-data\") pod \"barbican-db-sync-2hhl9\" (UID: \"c3b8b961-3c0e-47dc-b28f-d55c9821ab34\") " pod="openstack/barbican-db-sync-2hhl9" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.185924 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b70a321-ffb7-429f-b825-4cd872ded9ff-scripts\") pod \"cinder-db-sync-jfrdg\" (UID: \"7b70a321-ffb7-429f-b825-4cd872ded9ff\") " pod="openstack/cinder-db-sync-jfrdg" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.186030 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b70a321-ffb7-429f-b825-4cd872ded9ff-config-data\") pod \"cinder-db-sync-jfrdg\" (UID: \"7b70a321-ffb7-429f-b825-4cd872ded9ff\") " pod="openstack/cinder-db-sync-jfrdg" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.186188 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b70a321-ffb7-429f-b825-4cd872ded9ff-combined-ca-bundle\") pod \"cinder-db-sync-jfrdg\" (UID: \"7b70a321-ffb7-429f-b825-4cd872ded9ff\") " pod="openstack/cinder-db-sync-jfrdg" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.186292 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7b70a321-ffb7-429f-b825-4cd872ded9ff-db-sync-config-data\") pod \"cinder-db-sync-jfrdg\" (UID: \"7b70a321-ffb7-429f-b825-4cd872ded9ff\") " pod="openstack/cinder-db-sync-jfrdg" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.186409 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hw67n\" (UniqueName: \"kubernetes.io/projected/c3b8b961-3c0e-47dc-b28f-d55c9821ab34-kube-api-access-hw67n\") pod \"barbican-db-sync-2hhl9\" (UID: \"c3b8b961-3c0e-47dc-b28f-d55c9821ab34\") " pod="openstack/barbican-db-sync-2hhl9" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.186611 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3b8b961-3c0e-47dc-b28f-d55c9821ab34-combined-ca-bundle\") pod \"barbican-db-sync-2hhl9\" (UID: \"c3b8b961-3c0e-47dc-b28f-d55c9821ab34\") " pod="openstack/barbican-db-sync-2hhl9" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.186785 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmzf9\" (UniqueName: \"kubernetes.io/projected/7b70a321-ffb7-429f-b825-4cd872ded9ff-kube-api-access-zmzf9\") pod \"cinder-db-sync-jfrdg\" (UID: \"7b70a321-ffb7-429f-b825-4cd872ded9ff\") " pod="openstack/cinder-db-sync-jfrdg" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.186932 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7b70a321-ffb7-429f-b825-4cd872ded9ff-etc-machine-id\") pod \"cinder-db-sync-jfrdg\" (UID: \"7b70a321-ffb7-429f-b825-4cd872ded9ff\") " pod="openstack/cinder-db-sync-jfrdg" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.187170 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7b70a321-ffb7-429f-b825-4cd872ded9ff-etc-machine-id\") pod \"cinder-db-sync-jfrdg\" (UID: \"7b70a321-ffb7-429f-b825-4cd872ded9ff\") " pod="openstack/cinder-db-sync-jfrdg" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.196707 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c3b8b961-3c0e-47dc-b28f-d55c9821ab34-db-sync-config-data\") pod \"barbican-db-sync-2hhl9\" (UID: \"c3b8b961-3c0e-47dc-b28f-d55c9821ab34\") " pod="openstack/barbican-db-sync-2hhl9" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.196917 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7b70a321-ffb7-429f-b825-4cd872ded9ff-db-sync-config-data\") pod \"cinder-db-sync-jfrdg\" (UID: \"7b70a321-ffb7-429f-b825-4cd872ded9ff\") " pod="openstack/cinder-db-sync-jfrdg" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.198929 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b70a321-ffb7-429f-b825-4cd872ded9ff-config-data\") pod \"cinder-db-sync-jfrdg\" (UID: \"7b70a321-ffb7-429f-b825-4cd872ded9ff\") " pod="openstack/cinder-db-sync-jfrdg" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.200581 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b70a321-ffb7-429f-b825-4cd872ded9ff-combined-ca-bundle\") pod \"cinder-db-sync-jfrdg\" (UID: \"7b70a321-ffb7-429f-b825-4cd872ded9ff\") " pod="openstack/cinder-db-sync-jfrdg" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.200883 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b70a321-ffb7-429f-b825-4cd872ded9ff-scripts\") pod \"cinder-db-sync-jfrdg\" (UID: \"7b70a321-ffb7-429f-b825-4cd872ded9ff\") " pod="openstack/cinder-db-sync-jfrdg" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.201546 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3b8b961-3c0e-47dc-b28f-d55c9821ab34-combined-ca-bundle\") pod \"barbican-db-sync-2hhl9\" (UID: \"c3b8b961-3c0e-47dc-b28f-d55c9821ab34\") " pod="openstack/barbican-db-sync-2hhl9" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.213129 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmzf9\" (UniqueName: \"kubernetes.io/projected/7b70a321-ffb7-429f-b825-4cd872ded9ff-kube-api-access-zmzf9\") pod \"cinder-db-sync-jfrdg\" (UID: \"7b70a321-ffb7-429f-b825-4cd872ded9ff\") " pod="openstack/cinder-db-sync-jfrdg" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.214801 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hw67n\" (UniqueName: \"kubernetes.io/projected/c3b8b961-3c0e-47dc-b28f-d55c9821ab34-kube-api-access-hw67n\") pod \"barbican-db-sync-2hhl9\" (UID: \"c3b8b961-3c0e-47dc-b28f-d55c9821ab34\") " pod="openstack/barbican-db-sync-2hhl9" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.266383 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2hhl9" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.288233 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjmfq\" (UniqueName: \"kubernetes.io/projected/806fce72-373c-4c34-8293-3051691eb55f-kube-api-access-hjmfq\") pod \"neutron-db-sync-4sfdr\" (UID: \"806fce72-373c-4c34-8293-3051691eb55f\") " pod="openstack/neutron-db-sync-4sfdr" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.288323 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/806fce72-373c-4c34-8293-3051691eb55f-combined-ca-bundle\") pod \"neutron-db-sync-4sfdr\" (UID: \"806fce72-373c-4c34-8293-3051691eb55f\") " pod="openstack/neutron-db-sync-4sfdr" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.288412 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/806fce72-373c-4c34-8293-3051691eb55f-config\") pod \"neutron-db-sync-4sfdr\" (UID: \"806fce72-373c-4c34-8293-3051691eb55f\") " pod="openstack/neutron-db-sync-4sfdr" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.299385 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-jfrdg" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.390649 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/806fce72-373c-4c34-8293-3051691eb55f-config\") pod \"neutron-db-sync-4sfdr\" (UID: \"806fce72-373c-4c34-8293-3051691eb55f\") " pod="openstack/neutron-db-sync-4sfdr" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.390723 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjmfq\" (UniqueName: \"kubernetes.io/projected/806fce72-373c-4c34-8293-3051691eb55f-kube-api-access-hjmfq\") pod \"neutron-db-sync-4sfdr\" (UID: \"806fce72-373c-4c34-8293-3051691eb55f\") " pod="openstack/neutron-db-sync-4sfdr" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.390790 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/806fce72-373c-4c34-8293-3051691eb55f-combined-ca-bundle\") pod \"neutron-db-sync-4sfdr\" (UID: \"806fce72-373c-4c34-8293-3051691eb55f\") " pod="openstack/neutron-db-sync-4sfdr" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.395243 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/806fce72-373c-4c34-8293-3051691eb55f-combined-ca-bundle\") pod \"neutron-db-sync-4sfdr\" (UID: \"806fce72-373c-4c34-8293-3051691eb55f\") " pod="openstack/neutron-db-sync-4sfdr" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.408299 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/806fce72-373c-4c34-8293-3051691eb55f-config\") pod \"neutron-db-sync-4sfdr\" (UID: \"806fce72-373c-4c34-8293-3051691eb55f\") " pod="openstack/neutron-db-sync-4sfdr" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.414958 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjmfq\" (UniqueName: \"kubernetes.io/projected/806fce72-373c-4c34-8293-3051691eb55f-kube-api-access-hjmfq\") pod \"neutron-db-sync-4sfdr\" (UID: \"806fce72-373c-4c34-8293-3051691eb55f\") " pod="openstack/neutron-db-sync-4sfdr" Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.572363 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4sfdr" Oct 07 14:02:36 crc kubenswrapper[4959]: W1007 14:02:36.743999 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3b8b961_3c0e_47dc_b28f_d55c9821ab34.slice/crio-9b40f4a3a03e61fc71e06571fa3dffb62281881bb7810b6d9fed9638f36c7a3d WatchSource:0}: Error finding container 9b40f4a3a03e61fc71e06571fa3dffb62281881bb7810b6d9fed9638f36c7a3d: Status 404 returned error can't find the container with id 9b40f4a3a03e61fc71e06571fa3dffb62281881bb7810b6d9fed9638f36c7a3d Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.753642 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-2hhl9"] Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.785653 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-748d7644cf-6zpv4" event={"ID":"0ff263db-2611-4310-a9ee-01e2136c8b1b","Type":"ContainerStarted","Data":"adc28831f6472e2be951a70170e0355baa89eb96c0882c12dc2684466616c423"} Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.788342 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2hhl9" event={"ID":"c3b8b961-3c0e-47dc-b28f-d55c9821ab34","Type":"ContainerStarted","Data":"9b40f4a3a03e61fc71e06571fa3dffb62281881bb7810b6d9fed9638f36c7a3d"} Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.789299 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f459969c7-55hf7" event={"ID":"594da4a6-8781-416f-abdc-8f694948b6a2","Type":"ContainerStarted","Data":"ef90acde1c02f78da6ee25dd4a23b978ff700b3a89c4af4bd1a7781af60004f0"} Oct 07 14:02:36 crc kubenswrapper[4959]: I1007 14:02:36.832368 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-jfrdg"] Oct 07 14:02:36 crc kubenswrapper[4959]: W1007 14:02:36.835580 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b70a321_ffb7_429f_b825_4cd872ded9ff.slice/crio-5770a8e64286b3347c919d50f8ccdd718c75c59f7e590a2b8c08aac30333c6ea WatchSource:0}: Error finding container 5770a8e64286b3347c919d50f8ccdd718c75c59f7e590a2b8c08aac30333c6ea: Status 404 returned error can't find the container with id 5770a8e64286b3347c919d50f8ccdd718c75c59f7e590a2b8c08aac30333c6ea Oct 07 14:02:37 crc kubenswrapper[4959]: I1007 14:02:37.032711 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-4sfdr"] Oct 07 14:02:37 crc kubenswrapper[4959]: W1007 14:02:37.040053 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod806fce72_373c_4c34_8293_3051691eb55f.slice/crio-51270a802483bad43e9e5b3f35f56ee3aaaa4587185b43816ee5320d94406ec0 WatchSource:0}: Error finding container 51270a802483bad43e9e5b3f35f56ee3aaaa4587185b43816ee5320d94406ec0: Status 404 returned error can't find the container with id 51270a802483bad43e9e5b3f35f56ee3aaaa4587185b43816ee5320d94406ec0 Oct 07 14:02:37 crc kubenswrapper[4959]: I1007 14:02:37.802545 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-jfrdg" event={"ID":"7b70a321-ffb7-429f-b825-4cd872ded9ff","Type":"ContainerStarted","Data":"5770a8e64286b3347c919d50f8ccdd718c75c59f7e590a2b8c08aac30333c6ea"} Oct 07 14:02:37 crc kubenswrapper[4959]: I1007 14:02:37.803929 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4sfdr" event={"ID":"806fce72-373c-4c34-8293-3051691eb55f","Type":"ContainerStarted","Data":"51270a802483bad43e9e5b3f35f56ee3aaaa4587185b43816ee5320d94406ec0"} Oct 07 14:02:39 crc kubenswrapper[4959]: I1007 14:02:39.855769 4959 generic.go:334] "Generic (PLEG): container finished" podID="ee96a602-a1b4-4463-a919-b90fa83e23d3" containerID="45f9b19a49f20e63a4103a1e4057305cd8d055ef274fe1ec139759a62af9461b" exitCode=0 Oct 07 14:02:39 crc kubenswrapper[4959]: I1007 14:02:39.856287 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dcb7bb4dc-zz698" event={"ID":"ee96a602-a1b4-4463-a919-b90fa83e23d3","Type":"ContainerDied","Data":"45f9b19a49f20e63a4103a1e4057305cd8d055ef274fe1ec139759a62af9461b"} Oct 07 14:02:39 crc kubenswrapper[4959]: I1007 14:02:39.860254 4959 generic.go:334] "Generic (PLEG): container finished" podID="553dac39-5e7d-4cd4-8f2d-e096814c34c7" containerID="cc80adeab716193bbdee11668a04b96b7fa17d30fbdf0ba8a0543c3afae77821" exitCode=0 Oct 07 14:02:39 crc kubenswrapper[4959]: I1007 14:02:39.860307 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f549f7b99-nhlx5" event={"ID":"553dac39-5e7d-4cd4-8f2d-e096814c34c7","Type":"ContainerDied","Data":"cc80adeab716193bbdee11668a04b96b7fa17d30fbdf0ba8a0543c3afae77821"} Oct 07 14:02:39 crc kubenswrapper[4959]: I1007 14:02:39.863642 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v6lrw" event={"ID":"3205491a-c78b-4eff-b2e1-aa17eb81f5b4","Type":"ContainerStarted","Data":"2718b09b1af576bdba5f8d95924e279615a246feeba4cd05f9a0e23a587642f1"} Oct 07 14:02:39 crc kubenswrapper[4959]: I1007 14:02:39.867603 4959 generic.go:334] "Generic (PLEG): container finished" podID="0ff263db-2611-4310-a9ee-01e2136c8b1b" containerID="5096fb136c7910c43dec70435c7f47767bafcf91a92c228f8436148cd995b837" exitCode=0 Oct 07 14:02:39 crc kubenswrapper[4959]: I1007 14:02:39.867693 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-748d7644cf-6zpv4" event={"ID":"0ff263db-2611-4310-a9ee-01e2136c8b1b","Type":"ContainerDied","Data":"5096fb136c7910c43dec70435c7f47767bafcf91a92c228f8436148cd995b837"} Oct 07 14:02:39 crc kubenswrapper[4959]: I1007 14:02:39.893700 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4sfdr" event={"ID":"806fce72-373c-4c34-8293-3051691eb55f","Type":"ContainerStarted","Data":"d0594f269876109a127823477b2fc4b883d1d61793068015b09be8b59b07c22a"} Oct 07 14:02:39 crc kubenswrapper[4959]: I1007 14:02:39.986827 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-v6lrw" podStartSLOduration=7.986806163 podStartE2EDuration="7.986806163s" podCreationTimestamp="2025-10-07 14:02:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:02:39.970713394 +0000 UTC m=+1042.054117719" watchObservedRunningTime="2025-10-07 14:02:39.986806163 +0000 UTC m=+1042.070210488" Oct 07 14:02:40 crc kubenswrapper[4959]: I1007 14:02:40.001064 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-4sfdr" podStartSLOduration=4.001049636 podStartE2EDuration="4.001049636s" podCreationTimestamp="2025-10-07 14:02:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:02:39.990388902 +0000 UTC m=+1042.073793227" watchObservedRunningTime="2025-10-07 14:02:40.001049636 +0000 UTC m=+1042.084453961" Oct 07 14:02:40 crc kubenswrapper[4959]: I1007 14:02:40.236896 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f549f7b99-nhlx5" Oct 07 14:02:40 crc kubenswrapper[4959]: I1007 14:02:40.374993 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/553dac39-5e7d-4cd4-8f2d-e096814c34c7-config\") pod \"553dac39-5e7d-4cd4-8f2d-e096814c34c7\" (UID: \"553dac39-5e7d-4cd4-8f2d-e096814c34c7\") " Oct 07 14:02:40 crc kubenswrapper[4959]: I1007 14:02:40.375604 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/553dac39-5e7d-4cd4-8f2d-e096814c34c7-ovsdbserver-nb\") pod \"553dac39-5e7d-4cd4-8f2d-e096814c34c7\" (UID: \"553dac39-5e7d-4cd4-8f2d-e096814c34c7\") " Oct 07 14:02:40 crc kubenswrapper[4959]: I1007 14:02:40.375654 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/553dac39-5e7d-4cd4-8f2d-e096814c34c7-dns-svc\") pod \"553dac39-5e7d-4cd4-8f2d-e096814c34c7\" (UID: \"553dac39-5e7d-4cd4-8f2d-e096814c34c7\") " Oct 07 14:02:40 crc kubenswrapper[4959]: I1007 14:02:40.375856 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/553dac39-5e7d-4cd4-8f2d-e096814c34c7-ovsdbserver-sb\") pod \"553dac39-5e7d-4cd4-8f2d-e096814c34c7\" (UID: \"553dac39-5e7d-4cd4-8f2d-e096814c34c7\") " Oct 07 14:02:40 crc kubenswrapper[4959]: I1007 14:02:40.375942 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8wxn\" (UniqueName: \"kubernetes.io/projected/553dac39-5e7d-4cd4-8f2d-e096814c34c7-kube-api-access-d8wxn\") pod \"553dac39-5e7d-4cd4-8f2d-e096814c34c7\" (UID: \"553dac39-5e7d-4cd4-8f2d-e096814c34c7\") " Oct 07 14:02:40 crc kubenswrapper[4959]: I1007 14:02:40.381876 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/553dac39-5e7d-4cd4-8f2d-e096814c34c7-kube-api-access-d8wxn" (OuterVolumeSpecName: "kube-api-access-d8wxn") pod "553dac39-5e7d-4cd4-8f2d-e096814c34c7" (UID: "553dac39-5e7d-4cd4-8f2d-e096814c34c7"). InnerVolumeSpecName "kube-api-access-d8wxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:02:40 crc kubenswrapper[4959]: I1007 14:02:40.408856 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/553dac39-5e7d-4cd4-8f2d-e096814c34c7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "553dac39-5e7d-4cd4-8f2d-e096814c34c7" (UID: "553dac39-5e7d-4cd4-8f2d-e096814c34c7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:02:40 crc kubenswrapper[4959]: I1007 14:02:40.415265 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/553dac39-5e7d-4cd4-8f2d-e096814c34c7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "553dac39-5e7d-4cd4-8f2d-e096814c34c7" (UID: "553dac39-5e7d-4cd4-8f2d-e096814c34c7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:02:40 crc kubenswrapper[4959]: I1007 14:02:40.417457 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/553dac39-5e7d-4cd4-8f2d-e096814c34c7-config" (OuterVolumeSpecName: "config") pod "553dac39-5e7d-4cd4-8f2d-e096814c34c7" (UID: "553dac39-5e7d-4cd4-8f2d-e096814c34c7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:02:40 crc kubenswrapper[4959]: I1007 14:02:40.418210 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/553dac39-5e7d-4cd4-8f2d-e096814c34c7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "553dac39-5e7d-4cd4-8f2d-e096814c34c7" (UID: "553dac39-5e7d-4cd4-8f2d-e096814c34c7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:02:40 crc kubenswrapper[4959]: I1007 14:02:40.478322 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/553dac39-5e7d-4cd4-8f2d-e096814c34c7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:40 crc kubenswrapper[4959]: I1007 14:02:40.478357 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/553dac39-5e7d-4cd4-8f2d-e096814c34c7-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:40 crc kubenswrapper[4959]: I1007 14:02:40.478367 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/553dac39-5e7d-4cd4-8f2d-e096814c34c7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:40 crc kubenswrapper[4959]: I1007 14:02:40.478377 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8wxn\" (UniqueName: \"kubernetes.io/projected/553dac39-5e7d-4cd4-8f2d-e096814c34c7-kube-api-access-d8wxn\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:40 crc kubenswrapper[4959]: I1007 14:02:40.478386 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/553dac39-5e7d-4cd4-8f2d-e096814c34c7-config\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:40 crc kubenswrapper[4959]: I1007 14:02:40.908783 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f549f7b99-nhlx5" event={"ID":"553dac39-5e7d-4cd4-8f2d-e096814c34c7","Type":"ContainerDied","Data":"2ea96f487963ecf53d1963a52fcce8fd3233c7f69f727aa6085a7bcfc276a13d"} Oct 07 14:02:40 crc kubenswrapper[4959]: I1007 14:02:40.908850 4959 scope.go:117] "RemoveContainer" containerID="cc80adeab716193bbdee11668a04b96b7fa17d30fbdf0ba8a0543c3afae77821" Oct 07 14:02:40 crc kubenswrapper[4959]: I1007 14:02:40.909009 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f549f7b99-nhlx5" Oct 07 14:02:40 crc kubenswrapper[4959]: I1007 14:02:40.914837 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-748d7644cf-6zpv4" event={"ID":"0ff263db-2611-4310-a9ee-01e2136c8b1b","Type":"ContainerStarted","Data":"a5d00af38560f9ea04487166d9f58ac15892c23992b3e960a2f097c414cc9f58"} Oct 07 14:02:40 crc kubenswrapper[4959]: I1007 14:02:40.941893 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-748d7644cf-6zpv4" podStartSLOduration=5.9418793690000005 podStartE2EDuration="5.941879369s" podCreationTimestamp="2025-10-07 14:02:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:02:40.937519301 +0000 UTC m=+1043.020923626" watchObservedRunningTime="2025-10-07 14:02:40.941879369 +0000 UTC m=+1043.025283694" Oct 07 14:02:40 crc kubenswrapper[4959]: I1007 14:02:40.997591 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f549f7b99-nhlx5"] Oct 07 14:02:41 crc kubenswrapper[4959]: I1007 14:02:41.008836 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f549f7b99-nhlx5"] Oct 07 14:02:41 crc kubenswrapper[4959]: I1007 14:02:41.110518 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dcb7bb4dc-zz698" Oct 07 14:02:41 crc kubenswrapper[4959]: I1007 14:02:41.299833 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee96a602-a1b4-4463-a919-b90fa83e23d3-config\") pod \"ee96a602-a1b4-4463-a919-b90fa83e23d3\" (UID: \"ee96a602-a1b4-4463-a919-b90fa83e23d3\") " Oct 07 14:02:41 crc kubenswrapper[4959]: I1007 14:02:41.299913 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee96a602-a1b4-4463-a919-b90fa83e23d3-ovsdbserver-nb\") pod \"ee96a602-a1b4-4463-a919-b90fa83e23d3\" (UID: \"ee96a602-a1b4-4463-a919-b90fa83e23d3\") " Oct 07 14:02:41 crc kubenswrapper[4959]: I1007 14:02:41.300013 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee96a602-a1b4-4463-a919-b90fa83e23d3-dns-svc\") pod \"ee96a602-a1b4-4463-a919-b90fa83e23d3\" (UID: \"ee96a602-a1b4-4463-a919-b90fa83e23d3\") " Oct 07 14:02:41 crc kubenswrapper[4959]: I1007 14:02:41.300043 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee96a602-a1b4-4463-a919-b90fa83e23d3-ovsdbserver-sb\") pod \"ee96a602-a1b4-4463-a919-b90fa83e23d3\" (UID: \"ee96a602-a1b4-4463-a919-b90fa83e23d3\") " Oct 07 14:02:41 crc kubenswrapper[4959]: I1007 14:02:41.300093 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2qvh\" (UniqueName: \"kubernetes.io/projected/ee96a602-a1b4-4463-a919-b90fa83e23d3-kube-api-access-x2qvh\") pod \"ee96a602-a1b4-4463-a919-b90fa83e23d3\" (UID: \"ee96a602-a1b4-4463-a919-b90fa83e23d3\") " Oct 07 14:02:41 crc kubenswrapper[4959]: I1007 14:02:41.314070 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee96a602-a1b4-4463-a919-b90fa83e23d3-kube-api-access-x2qvh" (OuterVolumeSpecName: "kube-api-access-x2qvh") pod "ee96a602-a1b4-4463-a919-b90fa83e23d3" (UID: "ee96a602-a1b4-4463-a919-b90fa83e23d3"). InnerVolumeSpecName "kube-api-access-x2qvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:02:41 crc kubenswrapper[4959]: I1007 14:02:41.329033 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee96a602-a1b4-4463-a919-b90fa83e23d3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ee96a602-a1b4-4463-a919-b90fa83e23d3" (UID: "ee96a602-a1b4-4463-a919-b90fa83e23d3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:02:41 crc kubenswrapper[4959]: I1007 14:02:41.348802 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee96a602-a1b4-4463-a919-b90fa83e23d3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ee96a602-a1b4-4463-a919-b90fa83e23d3" (UID: "ee96a602-a1b4-4463-a919-b90fa83e23d3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:02:41 crc kubenswrapper[4959]: I1007 14:02:41.349015 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee96a602-a1b4-4463-a919-b90fa83e23d3-config" (OuterVolumeSpecName: "config") pod "ee96a602-a1b4-4463-a919-b90fa83e23d3" (UID: "ee96a602-a1b4-4463-a919-b90fa83e23d3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:02:41 crc kubenswrapper[4959]: I1007 14:02:41.358743 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee96a602-a1b4-4463-a919-b90fa83e23d3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ee96a602-a1b4-4463-a919-b90fa83e23d3" (UID: "ee96a602-a1b4-4463-a919-b90fa83e23d3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:02:41 crc kubenswrapper[4959]: I1007 14:02:41.403190 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee96a602-a1b4-4463-a919-b90fa83e23d3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:41 crc kubenswrapper[4959]: I1007 14:02:41.403221 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2qvh\" (UniqueName: \"kubernetes.io/projected/ee96a602-a1b4-4463-a919-b90fa83e23d3-kube-api-access-x2qvh\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:41 crc kubenswrapper[4959]: I1007 14:02:41.403232 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee96a602-a1b4-4463-a919-b90fa83e23d3-config\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:41 crc kubenswrapper[4959]: I1007 14:02:41.403241 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee96a602-a1b4-4463-a919-b90fa83e23d3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:41 crc kubenswrapper[4959]: I1007 14:02:41.403248 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee96a602-a1b4-4463-a919-b90fa83e23d3-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 14:02:41 crc kubenswrapper[4959]: I1007 14:02:41.923948 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dcb7bb4dc-zz698" event={"ID":"ee96a602-a1b4-4463-a919-b90fa83e23d3","Type":"ContainerDied","Data":"4dfb81a288109f1aa6ac4cfb60e430ffe49364b0c3d4de9a8c7f7c671b3df005"} Oct 07 14:02:41 crc kubenswrapper[4959]: I1007 14:02:41.924369 4959 scope.go:117] "RemoveContainer" containerID="45f9b19a49f20e63a4103a1e4057305cd8d055ef274fe1ec139759a62af9461b" Oct 07 14:02:41 crc kubenswrapper[4959]: I1007 14:02:41.924451 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dcb7bb4dc-zz698" Oct 07 14:02:41 crc kubenswrapper[4959]: I1007 14:02:41.937466 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-748d7644cf-6zpv4" Oct 07 14:02:41 crc kubenswrapper[4959]: I1007 14:02:41.996433 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dcb7bb4dc-zz698"] Oct 07 14:02:42 crc kubenswrapper[4959]: I1007 14:02:42.000408 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5dcb7bb4dc-zz698"] Oct 07 14:02:42 crc kubenswrapper[4959]: I1007 14:02:42.664210 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="553dac39-5e7d-4cd4-8f2d-e096814c34c7" path="/var/lib/kubelet/pods/553dac39-5e7d-4cd4-8f2d-e096814c34c7/volumes" Oct 07 14:02:42 crc kubenswrapper[4959]: I1007 14:02:42.664796 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee96a602-a1b4-4463-a919-b90fa83e23d3" path="/var/lib/kubelet/pods/ee96a602-a1b4-4463-a919-b90fa83e23d3/volumes" Oct 07 14:02:42 crc kubenswrapper[4959]: I1007 14:02:42.961608 4959 generic.go:334] "Generic (PLEG): container finished" podID="3205491a-c78b-4eff-b2e1-aa17eb81f5b4" containerID="2718b09b1af576bdba5f8d95924e279615a246feeba4cd05f9a0e23a587642f1" exitCode=0 Oct 07 14:02:42 crc kubenswrapper[4959]: I1007 14:02:42.961662 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v6lrw" event={"ID":"3205491a-c78b-4eff-b2e1-aa17eb81f5b4","Type":"ContainerDied","Data":"2718b09b1af576bdba5f8d95924e279615a246feeba4cd05f9a0e23a587642f1"} Oct 07 14:02:45 crc kubenswrapper[4959]: I1007 14:02:45.459286 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-748d7644cf-6zpv4" Oct 07 14:02:45 crc kubenswrapper[4959]: I1007 14:02:45.508244 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bc45f6dcf-t6p7q"] Oct 07 14:02:45 crc kubenswrapper[4959]: I1007 14:02:45.508467 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" podUID="4c502b70-d657-47d6-a703-4ecf3cc84d01" containerName="dnsmasq-dns" containerID="cri-o://a8a2200da57e5e47c36da78d602f44fd8a60bcac7cad79714ee6b62fcf3956d8" gracePeriod=10 Oct 07 14:02:45 crc kubenswrapper[4959]: I1007 14:02:45.859084 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" podUID="4c502b70-d657-47d6-a703-4ecf3cc84d01" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: connect: connection refused" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.081686 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-d5f5b8fd9-mdlh9"] Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.117737 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-744bf8f46d-6twnn"] Oct 07 14:02:46 crc kubenswrapper[4959]: E1007 14:02:46.118067 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee96a602-a1b4-4463-a919-b90fa83e23d3" containerName="init" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.118082 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee96a602-a1b4-4463-a919-b90fa83e23d3" containerName="init" Oct 07 14:02:46 crc kubenswrapper[4959]: E1007 14:02:46.118129 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="553dac39-5e7d-4cd4-8f2d-e096814c34c7" containerName="init" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.118135 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="553dac39-5e7d-4cd4-8f2d-e096814c34c7" containerName="init" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.118289 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="553dac39-5e7d-4cd4-8f2d-e096814c34c7" containerName="init" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.118309 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee96a602-a1b4-4463-a919-b90fa83e23d3" containerName="init" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.119071 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.120954 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.136395 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-744bf8f46d-6twnn"] Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.210606 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a588fbdc-fd65-49f6-ab59-4901c7ca690a-scripts\") pod \"horizon-744bf8f46d-6twnn\" (UID: \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\") " pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.210691 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a588fbdc-fd65-49f6-ab59-4901c7ca690a-config-data\") pod \"horizon-744bf8f46d-6twnn\" (UID: \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\") " pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.210724 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a588fbdc-fd65-49f6-ab59-4901c7ca690a-horizon-tls-certs\") pod \"horizon-744bf8f46d-6twnn\" (UID: \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\") " pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.210761 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a588fbdc-fd65-49f6-ab59-4901c7ca690a-horizon-secret-key\") pod \"horizon-744bf8f46d-6twnn\" (UID: \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\") " pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.210779 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a588fbdc-fd65-49f6-ab59-4901c7ca690a-logs\") pod \"horizon-744bf8f46d-6twnn\" (UID: \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\") " pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.210820 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9nlx\" (UniqueName: \"kubernetes.io/projected/a588fbdc-fd65-49f6-ab59-4901c7ca690a-kube-api-access-x9nlx\") pod \"horizon-744bf8f46d-6twnn\" (UID: \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\") " pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.210871 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a588fbdc-fd65-49f6-ab59-4901c7ca690a-combined-ca-bundle\") pod \"horizon-744bf8f46d-6twnn\" (UID: \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\") " pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.241850 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5f459969c7-55hf7"] Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.276701 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-76d6fc6cbd-2g7sq"] Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.278349 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-76d6fc6cbd-2g7sq" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.297136 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-76d6fc6cbd-2g7sq"] Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.315024 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a588fbdc-fd65-49f6-ab59-4901c7ca690a-combined-ca-bundle\") pod \"horizon-744bf8f46d-6twnn\" (UID: \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\") " pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.315119 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/13b27eed-f50f-4474-ace5-0f12e733f6cf-horizon-secret-key\") pod \"horizon-76d6fc6cbd-2g7sq\" (UID: \"13b27eed-f50f-4474-ace5-0f12e733f6cf\") " pod="openstack/horizon-76d6fc6cbd-2g7sq" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.315184 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13b27eed-f50f-4474-ace5-0f12e733f6cf-combined-ca-bundle\") pod \"horizon-76d6fc6cbd-2g7sq\" (UID: \"13b27eed-f50f-4474-ace5-0f12e733f6cf\") " pod="openstack/horizon-76d6fc6cbd-2g7sq" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.315250 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gz95f\" (UniqueName: \"kubernetes.io/projected/13b27eed-f50f-4474-ace5-0f12e733f6cf-kube-api-access-gz95f\") pod \"horizon-76d6fc6cbd-2g7sq\" (UID: \"13b27eed-f50f-4474-ace5-0f12e733f6cf\") " pod="openstack/horizon-76d6fc6cbd-2g7sq" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.315284 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a588fbdc-fd65-49f6-ab59-4901c7ca690a-scripts\") pod \"horizon-744bf8f46d-6twnn\" (UID: \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\") " pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.315318 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a588fbdc-fd65-49f6-ab59-4901c7ca690a-config-data\") pod \"horizon-744bf8f46d-6twnn\" (UID: \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\") " pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.315358 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a588fbdc-fd65-49f6-ab59-4901c7ca690a-horizon-tls-certs\") pod \"horizon-744bf8f46d-6twnn\" (UID: \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\") " pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.315394 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13b27eed-f50f-4474-ace5-0f12e733f6cf-logs\") pod \"horizon-76d6fc6cbd-2g7sq\" (UID: \"13b27eed-f50f-4474-ace5-0f12e733f6cf\") " pod="openstack/horizon-76d6fc6cbd-2g7sq" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.315418 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a588fbdc-fd65-49f6-ab59-4901c7ca690a-logs\") pod \"horizon-744bf8f46d-6twnn\" (UID: \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\") " pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.315443 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a588fbdc-fd65-49f6-ab59-4901c7ca690a-horizon-secret-key\") pod \"horizon-744bf8f46d-6twnn\" (UID: \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\") " pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.315489 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/13b27eed-f50f-4474-ace5-0f12e733f6cf-config-data\") pod \"horizon-76d6fc6cbd-2g7sq\" (UID: \"13b27eed-f50f-4474-ace5-0f12e733f6cf\") " pod="openstack/horizon-76d6fc6cbd-2g7sq" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.315519 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9nlx\" (UniqueName: \"kubernetes.io/projected/a588fbdc-fd65-49f6-ab59-4901c7ca690a-kube-api-access-x9nlx\") pod \"horizon-744bf8f46d-6twnn\" (UID: \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\") " pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.315553 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/13b27eed-f50f-4474-ace5-0f12e733f6cf-scripts\") pod \"horizon-76d6fc6cbd-2g7sq\" (UID: \"13b27eed-f50f-4474-ace5-0f12e733f6cf\") " pod="openstack/horizon-76d6fc6cbd-2g7sq" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.315583 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/13b27eed-f50f-4474-ace5-0f12e733f6cf-horizon-tls-certs\") pod \"horizon-76d6fc6cbd-2g7sq\" (UID: \"13b27eed-f50f-4474-ace5-0f12e733f6cf\") " pod="openstack/horizon-76d6fc6cbd-2g7sq" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.321640 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a588fbdc-fd65-49f6-ab59-4901c7ca690a-logs\") pod \"horizon-744bf8f46d-6twnn\" (UID: \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\") " pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.322225 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a588fbdc-fd65-49f6-ab59-4901c7ca690a-scripts\") pod \"horizon-744bf8f46d-6twnn\" (UID: \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\") " pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.323031 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a588fbdc-fd65-49f6-ab59-4901c7ca690a-config-data\") pod \"horizon-744bf8f46d-6twnn\" (UID: \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\") " pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.327636 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a588fbdc-fd65-49f6-ab59-4901c7ca690a-horizon-secret-key\") pod \"horizon-744bf8f46d-6twnn\" (UID: \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\") " pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.333652 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a588fbdc-fd65-49f6-ab59-4901c7ca690a-horizon-tls-certs\") pod \"horizon-744bf8f46d-6twnn\" (UID: \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\") " pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.339965 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a588fbdc-fd65-49f6-ab59-4901c7ca690a-combined-ca-bundle\") pod \"horizon-744bf8f46d-6twnn\" (UID: \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\") " pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.352628 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9nlx\" (UniqueName: \"kubernetes.io/projected/a588fbdc-fd65-49f6-ab59-4901c7ca690a-kube-api-access-x9nlx\") pod \"horizon-744bf8f46d-6twnn\" (UID: \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\") " pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.417476 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gz95f\" (UniqueName: \"kubernetes.io/projected/13b27eed-f50f-4474-ace5-0f12e733f6cf-kube-api-access-gz95f\") pod \"horizon-76d6fc6cbd-2g7sq\" (UID: \"13b27eed-f50f-4474-ace5-0f12e733f6cf\") " pod="openstack/horizon-76d6fc6cbd-2g7sq" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.417729 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13b27eed-f50f-4474-ace5-0f12e733f6cf-logs\") pod \"horizon-76d6fc6cbd-2g7sq\" (UID: \"13b27eed-f50f-4474-ace5-0f12e733f6cf\") " pod="openstack/horizon-76d6fc6cbd-2g7sq" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.417863 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/13b27eed-f50f-4474-ace5-0f12e733f6cf-config-data\") pod \"horizon-76d6fc6cbd-2g7sq\" (UID: \"13b27eed-f50f-4474-ace5-0f12e733f6cf\") " pod="openstack/horizon-76d6fc6cbd-2g7sq" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.417969 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/13b27eed-f50f-4474-ace5-0f12e733f6cf-scripts\") pod \"horizon-76d6fc6cbd-2g7sq\" (UID: \"13b27eed-f50f-4474-ace5-0f12e733f6cf\") " pod="openstack/horizon-76d6fc6cbd-2g7sq" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.418067 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/13b27eed-f50f-4474-ace5-0f12e733f6cf-horizon-tls-certs\") pod \"horizon-76d6fc6cbd-2g7sq\" (UID: \"13b27eed-f50f-4474-ace5-0f12e733f6cf\") " pod="openstack/horizon-76d6fc6cbd-2g7sq" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.418231 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/13b27eed-f50f-4474-ace5-0f12e733f6cf-horizon-secret-key\") pod \"horizon-76d6fc6cbd-2g7sq\" (UID: \"13b27eed-f50f-4474-ace5-0f12e733f6cf\") " pod="openstack/horizon-76d6fc6cbd-2g7sq" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.418462 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13b27eed-f50f-4474-ace5-0f12e733f6cf-combined-ca-bundle\") pod \"horizon-76d6fc6cbd-2g7sq\" (UID: \"13b27eed-f50f-4474-ace5-0f12e733f6cf\") " pod="openstack/horizon-76d6fc6cbd-2g7sq" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.418295 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13b27eed-f50f-4474-ace5-0f12e733f6cf-logs\") pod \"horizon-76d6fc6cbd-2g7sq\" (UID: \"13b27eed-f50f-4474-ace5-0f12e733f6cf\") " pod="openstack/horizon-76d6fc6cbd-2g7sq" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.419002 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/13b27eed-f50f-4474-ace5-0f12e733f6cf-scripts\") pod \"horizon-76d6fc6cbd-2g7sq\" (UID: \"13b27eed-f50f-4474-ace5-0f12e733f6cf\") " pod="openstack/horizon-76d6fc6cbd-2g7sq" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.419885 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/13b27eed-f50f-4474-ace5-0f12e733f6cf-config-data\") pod \"horizon-76d6fc6cbd-2g7sq\" (UID: \"13b27eed-f50f-4474-ace5-0f12e733f6cf\") " pod="openstack/horizon-76d6fc6cbd-2g7sq" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.422556 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/13b27eed-f50f-4474-ace5-0f12e733f6cf-horizon-secret-key\") pod \"horizon-76d6fc6cbd-2g7sq\" (UID: \"13b27eed-f50f-4474-ace5-0f12e733f6cf\") " pod="openstack/horizon-76d6fc6cbd-2g7sq" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.423160 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13b27eed-f50f-4474-ace5-0f12e733f6cf-combined-ca-bundle\") pod \"horizon-76d6fc6cbd-2g7sq\" (UID: \"13b27eed-f50f-4474-ace5-0f12e733f6cf\") " pod="openstack/horizon-76d6fc6cbd-2g7sq" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.425744 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/13b27eed-f50f-4474-ace5-0f12e733f6cf-horizon-tls-certs\") pod \"horizon-76d6fc6cbd-2g7sq\" (UID: \"13b27eed-f50f-4474-ace5-0f12e733f6cf\") " pod="openstack/horizon-76d6fc6cbd-2g7sq" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.433145 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gz95f\" (UniqueName: \"kubernetes.io/projected/13b27eed-f50f-4474-ace5-0f12e733f6cf-kube-api-access-gz95f\") pod \"horizon-76d6fc6cbd-2g7sq\" (UID: \"13b27eed-f50f-4474-ace5-0f12e733f6cf\") " pod="openstack/horizon-76d6fc6cbd-2g7sq" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.486950 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:02:46 crc kubenswrapper[4959]: I1007 14:02:46.608981 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-76d6fc6cbd-2g7sq" Oct 07 14:02:47 crc kubenswrapper[4959]: I1007 14:02:47.010782 4959 generic.go:334] "Generic (PLEG): container finished" podID="4c502b70-d657-47d6-a703-4ecf3cc84d01" containerID="a8a2200da57e5e47c36da78d602f44fd8a60bcac7cad79714ee6b62fcf3956d8" exitCode=0 Oct 07 14:02:47 crc kubenswrapper[4959]: I1007 14:02:47.010827 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" event={"ID":"4c502b70-d657-47d6-a703-4ecf3cc84d01","Type":"ContainerDied","Data":"a8a2200da57e5e47c36da78d602f44fd8a60bcac7cad79714ee6b62fcf3956d8"} Oct 07 14:02:50 crc kubenswrapper[4959]: I1007 14:02:50.858961 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" podUID="4c502b70-d657-47d6-a703-4ecf3cc84d01" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: connect: connection refused" Oct 07 14:02:55 crc kubenswrapper[4959]: E1007 14:02:55.592964 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon@sha256:03b4bb79b71d5ca7792d19c4c0ee08a5e5a407ad844c087305c42dd909ee7490" Oct 07 14:02:55 crc kubenswrapper[4959]: E1007 14:02:55.593747 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon@sha256:03b4bb79b71d5ca7792d19c4c0ee08a5e5a407ad844c087305c42dd909ee7490,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nb8h5cch584h5d9h7bhf4h5cdh5ffh55bh69h56bh677hc6h65dh57dh674h66dh5fdh689h5fdh68bh5c4h66chbch598h57dh5b5h68fh65ch548h7ch5dq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jzcc6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-59bc478f75-xdfhn_openstack(5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 14:02:55 crc kubenswrapper[4959]: E1007 14:02:55.598496 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon@sha256:03b4bb79b71d5ca7792d19c4c0ee08a5e5a407ad844c087305c42dd909ee7490\\\"\"]" pod="openstack/horizon-59bc478f75-xdfhn" podUID="5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0" Oct 07 14:02:55 crc kubenswrapper[4959]: I1007 14:02:55.859270 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" podUID="4c502b70-d657-47d6-a703-4ecf3cc84d01" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: connect: connection refused" Oct 07 14:02:55 crc kubenswrapper[4959]: I1007 14:02:55.859888 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" Oct 07 14:02:56 crc kubenswrapper[4959]: E1007 14:02:56.634816 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:cbe345acb37e57986ecf6685d28c72d0e639bdb493a18e9d3ba947d6c3a16384" Oct 07 14:02:56 crc kubenswrapper[4959]: E1007 14:02:56.635362 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:cbe345acb37e57986ecf6685d28c72d0e639bdb493a18e9d3ba947d6c3a16384,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hw67n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-2hhl9_openstack(c3b8b961-3c0e-47dc-b28f-d55c9821ab34): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 14:02:56 crc kubenswrapper[4959]: E1007 14:02:56.638543 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-2hhl9" podUID="c3b8b961-3c0e-47dc-b28f-d55c9821ab34" Oct 07 14:02:56 crc kubenswrapper[4959]: E1007 14:02:56.771335 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon@sha256:03b4bb79b71d5ca7792d19c4c0ee08a5e5a407ad844c087305c42dd909ee7490" Oct 07 14:02:56 crc kubenswrapper[4959]: E1007 14:02:56.771587 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon@sha256:03b4bb79b71d5ca7792d19c4c0ee08a5e5a407ad844c087305c42dd909ee7490,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n59h58h6bh5dbh566hf5h55h7bh555h64fh77h84h58bh565hf9hf8h58fh697h64ch696h56hf7h99h575hfchbch677h645h556h5bbh5d5h547q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c6nvt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-d5f5b8fd9-mdlh9_openstack(906180a5-7c20-4109-8094-7d578d8ad374): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 14:02:56 crc kubenswrapper[4959]: E1007 14:02:56.774370 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon@sha256:03b4bb79b71d5ca7792d19c4c0ee08a5e5a407ad844c087305c42dd909ee7490\\\"\"]" pod="openstack/horizon-d5f5b8fd9-mdlh9" podUID="906180a5-7c20-4109-8094-7d578d8ad374" Oct 07 14:02:56 crc kubenswrapper[4959]: E1007 14:02:56.796793 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon@sha256:03b4bb79b71d5ca7792d19c4c0ee08a5e5a407ad844c087305c42dd909ee7490" Oct 07 14:02:56 crc kubenswrapper[4959]: E1007 14:02:56.796943 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon@sha256:03b4bb79b71d5ca7792d19c4c0ee08a5e5a407ad844c087305c42dd909ee7490,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n64dh54bh568h5f6h88h88hb5h58dh5c8h57ch59ch597h5c7h7h5bch677h654h67h5bfh54bhf6h55fh57bh5dbh578h65bh5cfh668h545h95hdh678q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7x4vd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-5f459969c7-55hf7_openstack(594da4a6-8781-416f-abdc-8f694948b6a2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 14:02:56 crc kubenswrapper[4959]: E1007 14:02:56.799150 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon@sha256:03b4bb79b71d5ca7792d19c4c0ee08a5e5a407ad844c087305c42dd909ee7490\\\"\"]" pod="openstack/horizon-5f459969c7-55hf7" podUID="594da4a6-8781-416f-abdc-8f694948b6a2" Oct 07 14:02:56 crc kubenswrapper[4959]: E1007 14:02:56.965065 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:86daeb9c834bfcedb533086dff59a6b5b6e832b94ce2a9116337f8736bb80032" Oct 07 14:02:56 crc kubenswrapper[4959]: E1007 14:02:56.965262 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:86daeb9c834bfcedb533086dff59a6b5b6e832b94ce2a9116337f8736bb80032,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n56fh5c7h5b5h65fh65fh64hffh66fhd5h98hd9h5b5h585h56fh665h5c6h8fh59h675h644hbdh5ffhcfh588h684h87h584h5dbh684h569h667h59bq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lcw9j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(cd33b52c-3f7f-425e-aa42-7342343067a6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 14:02:57 crc kubenswrapper[4959]: E1007 14:02:57.090341 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:cbe345acb37e57986ecf6685d28c72d0e639bdb493a18e9d3ba947d6c3a16384\\\"\"" pod="openstack/barbican-db-sync-2hhl9" podUID="c3b8b961-3c0e-47dc-b28f-d55c9821ab34" Oct 07 14:03:00 crc kubenswrapper[4959]: I1007 14:03:00.859903 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" podUID="4c502b70-d657-47d6-a703-4ecf3cc84d01" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: connect: connection refused" Oct 07 14:03:05 crc kubenswrapper[4959]: I1007 14:03:05.858957 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" podUID="4c502b70-d657-47d6-a703-4ecf3cc84d01" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: connect: connection refused" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.451340 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v6lrw" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.458497 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-59bc478f75-xdfhn" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.466815 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d5f5b8fd9-mdlh9" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.491249 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f459969c7-55hf7" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.550749 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-combined-ca-bundle\") pod \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\" (UID: \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\") " Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.551020 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-logs\") pod \"5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0\" (UID: \"5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0\") " Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.551047 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/906180a5-7c20-4109-8094-7d578d8ad374-scripts\") pod \"906180a5-7c20-4109-8094-7d578d8ad374\" (UID: \"906180a5-7c20-4109-8094-7d578d8ad374\") " Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.551138 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/906180a5-7c20-4109-8094-7d578d8ad374-horizon-secret-key\") pod \"906180a5-7c20-4109-8094-7d578d8ad374\" (UID: \"906180a5-7c20-4109-8094-7d578d8ad374\") " Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.551181 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-scripts\") pod \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\" (UID: \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\") " Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.551210 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-config-data\") pod \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\" (UID: \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\") " Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.551251 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-fernet-keys\") pod \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\" (UID: \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\") " Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.551282 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/906180a5-7c20-4109-8094-7d578d8ad374-logs\") pod \"906180a5-7c20-4109-8094-7d578d8ad374\" (UID: \"906180a5-7c20-4109-8094-7d578d8ad374\") " Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.551337 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgjcq\" (UniqueName: \"kubernetes.io/projected/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-kube-api-access-fgjcq\") pod \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\" (UID: \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\") " Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.551362 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-horizon-secret-key\") pod \"5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0\" (UID: \"5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0\") " Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.551410 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6nvt\" (UniqueName: \"kubernetes.io/projected/906180a5-7c20-4109-8094-7d578d8ad374-kube-api-access-c6nvt\") pod \"906180a5-7c20-4109-8094-7d578d8ad374\" (UID: \"906180a5-7c20-4109-8094-7d578d8ad374\") " Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.551434 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/906180a5-7c20-4109-8094-7d578d8ad374-config-data\") pod \"906180a5-7c20-4109-8094-7d578d8ad374\" (UID: \"906180a5-7c20-4109-8094-7d578d8ad374\") " Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.551472 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-config-data\") pod \"5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0\" (UID: \"5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0\") " Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.551479 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-logs" (OuterVolumeSpecName: "logs") pod "5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0" (UID: "5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.551493 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzcc6\" (UniqueName: \"kubernetes.io/projected/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-kube-api-access-jzcc6\") pod \"5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0\" (UID: \"5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0\") " Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.551571 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-scripts\") pod \"5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0\" (UID: \"5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0\") " Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.551598 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-credential-keys\") pod \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\" (UID: \"3205491a-c78b-4eff-b2e1-aa17eb81f5b4\") " Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.551610 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/906180a5-7c20-4109-8094-7d578d8ad374-scripts" (OuterVolumeSpecName: "scripts") pod "906180a5-7c20-4109-8094-7d578d8ad374" (UID: "906180a5-7c20-4109-8094-7d578d8ad374"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.552370 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-logs\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.552388 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/906180a5-7c20-4109-8094-7d578d8ad374-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.557180 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/906180a5-7c20-4109-8094-7d578d8ad374-logs" (OuterVolumeSpecName: "logs") pod "906180a5-7c20-4109-8094-7d578d8ad374" (UID: "906180a5-7c20-4109-8094-7d578d8ad374"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.557761 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-kube-api-access-fgjcq" (OuterVolumeSpecName: "kube-api-access-fgjcq") pod "3205491a-c78b-4eff-b2e1-aa17eb81f5b4" (UID: "3205491a-c78b-4eff-b2e1-aa17eb81f5b4"). InnerVolumeSpecName "kube-api-access-fgjcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.558258 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-config-data" (OuterVolumeSpecName: "config-data") pod "5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0" (UID: "5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.559363 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/906180a5-7c20-4109-8094-7d578d8ad374-config-data" (OuterVolumeSpecName: "config-data") pod "906180a5-7c20-4109-8094-7d578d8ad374" (UID: "906180a5-7c20-4109-8094-7d578d8ad374"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.561436 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "3205491a-c78b-4eff-b2e1-aa17eb81f5b4" (UID: "3205491a-c78b-4eff-b2e1-aa17eb81f5b4"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.561491 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-scripts" (OuterVolumeSpecName: "scripts") pod "5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0" (UID: "5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.561813 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-kube-api-access-jzcc6" (OuterVolumeSpecName: "kube-api-access-jzcc6") pod "5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0" (UID: "5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0"). InnerVolumeSpecName "kube-api-access-jzcc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.562017 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/906180a5-7c20-4109-8094-7d578d8ad374-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "906180a5-7c20-4109-8094-7d578d8ad374" (UID: "906180a5-7c20-4109-8094-7d578d8ad374"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.564434 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/906180a5-7c20-4109-8094-7d578d8ad374-kube-api-access-c6nvt" (OuterVolumeSpecName: "kube-api-access-c6nvt") pod "906180a5-7c20-4109-8094-7d578d8ad374" (UID: "906180a5-7c20-4109-8094-7d578d8ad374"). InnerVolumeSpecName "kube-api-access-c6nvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.564989 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0" (UID: "5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.566663 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "3205491a-c78b-4eff-b2e1-aa17eb81f5b4" (UID: "3205491a-c78b-4eff-b2e1-aa17eb81f5b4"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.566698 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-scripts" (OuterVolumeSpecName: "scripts") pod "3205491a-c78b-4eff-b2e1-aa17eb81f5b4" (UID: "3205491a-c78b-4eff-b2e1-aa17eb81f5b4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.579061 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3205491a-c78b-4eff-b2e1-aa17eb81f5b4" (UID: "3205491a-c78b-4eff-b2e1-aa17eb81f5b4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.603480 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-config-data" (OuterVolumeSpecName: "config-data") pod "3205491a-c78b-4eff-b2e1-aa17eb81f5b4" (UID: "3205491a-c78b-4eff-b2e1-aa17eb81f5b4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.653451 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7x4vd\" (UniqueName: \"kubernetes.io/projected/594da4a6-8781-416f-abdc-8f694948b6a2-kube-api-access-7x4vd\") pod \"594da4a6-8781-416f-abdc-8f694948b6a2\" (UID: \"594da4a6-8781-416f-abdc-8f694948b6a2\") " Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.653564 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/594da4a6-8781-416f-abdc-8f694948b6a2-scripts\") pod \"594da4a6-8781-416f-abdc-8f694948b6a2\" (UID: \"594da4a6-8781-416f-abdc-8f694948b6a2\") " Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.653597 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/594da4a6-8781-416f-abdc-8f694948b6a2-horizon-secret-key\") pod \"594da4a6-8781-416f-abdc-8f694948b6a2\" (UID: \"594da4a6-8781-416f-abdc-8f694948b6a2\") " Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.653654 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/594da4a6-8781-416f-abdc-8f694948b6a2-logs\") pod \"594da4a6-8781-416f-abdc-8f694948b6a2\" (UID: \"594da4a6-8781-416f-abdc-8f694948b6a2\") " Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.654267 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/594da4a6-8781-416f-abdc-8f694948b6a2-logs" (OuterVolumeSpecName: "logs") pod "594da4a6-8781-416f-abdc-8f694948b6a2" (UID: "594da4a6-8781-416f-abdc-8f694948b6a2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.654292 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/594da4a6-8781-416f-abdc-8f694948b6a2-scripts" (OuterVolumeSpecName: "scripts") pod "594da4a6-8781-416f-abdc-8f694948b6a2" (UID: "594da4a6-8781-416f-abdc-8f694948b6a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.654412 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/594da4a6-8781-416f-abdc-8f694948b6a2-config-data\") pod \"594da4a6-8781-416f-abdc-8f694948b6a2\" (UID: \"594da4a6-8781-416f-abdc-8f694948b6a2\") " Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.654951 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/594da4a6-8781-416f-abdc-8f694948b6a2-config-data" (OuterVolumeSpecName: "config-data") pod "594da4a6-8781-416f-abdc-8f694948b6a2" (UID: "594da4a6-8781-416f-abdc-8f694948b6a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.655045 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.655186 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/594da4a6-8781-416f-abdc-8f694948b6a2-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.655215 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/594da4a6-8781-416f-abdc-8f694948b6a2-logs\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.655230 4959 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/906180a5-7c20-4109-8094-7d578d8ad374-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.655247 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.655359 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.655380 4959 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.655391 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/906180a5-7c20-4109-8094-7d578d8ad374-logs\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.655418 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgjcq\" (UniqueName: \"kubernetes.io/projected/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-kube-api-access-fgjcq\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.655431 4959 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.655443 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6nvt\" (UniqueName: \"kubernetes.io/projected/906180a5-7c20-4109-8094-7d578d8ad374-kube-api-access-c6nvt\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.655495 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/906180a5-7c20-4109-8094-7d578d8ad374-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.655513 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.655525 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzcc6\" (UniqueName: \"kubernetes.io/projected/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-kube-api-access-jzcc6\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.655537 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.655574 4959 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3205491a-c78b-4eff-b2e1-aa17eb81f5b4-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.656617 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/594da4a6-8781-416f-abdc-8f694948b6a2-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "594da4a6-8781-416f-abdc-8f694948b6a2" (UID: "594da4a6-8781-416f-abdc-8f694948b6a2"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.658898 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/594da4a6-8781-416f-abdc-8f694948b6a2-kube-api-access-7x4vd" (OuterVolumeSpecName: "kube-api-access-7x4vd") pod "594da4a6-8781-416f-abdc-8f694948b6a2" (UID: "594da4a6-8781-416f-abdc-8f694948b6a2"). InnerVolumeSpecName "kube-api-access-7x4vd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.757470 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/594da4a6-8781-416f-abdc-8f694948b6a2-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.757510 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7x4vd\" (UniqueName: \"kubernetes.io/projected/594da4a6-8781-416f-abdc-8f694948b6a2-kube-api-access-7x4vd\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:06 crc kubenswrapper[4959]: I1007 14:03:06.757526 4959 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/594da4a6-8781-416f-abdc-8f694948b6a2-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.197963 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v6lrw" event={"ID":"3205491a-c78b-4eff-b2e1-aa17eb81f5b4","Type":"ContainerDied","Data":"f635c32fa93ab50f3a6dcf3a67212f60c73f8dadab7fd02cd593a5d8f254d24b"} Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.198011 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f635c32fa93ab50f3a6dcf3a67212f60c73f8dadab7fd02cd593a5d8f254d24b" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.198117 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v6lrw" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.201469 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d5f5b8fd9-mdlh9" event={"ID":"906180a5-7c20-4109-8094-7d578d8ad374","Type":"ContainerDied","Data":"21cddbd5537b4e7e16caebd3f512d0d4aac613cb8cc0b2a4abd593bb11f8e3a2"} Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.201516 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d5f5b8fd9-mdlh9" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.204209 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-59bc478f75-xdfhn" event={"ID":"5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0","Type":"ContainerDied","Data":"3e92680d8b5a23c3bae9589d5d17b34b82712d00baf737b9f3d1445ec475e0c0"} Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.204291 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-59bc478f75-xdfhn" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.206311 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f459969c7-55hf7" event={"ID":"594da4a6-8781-416f-abdc-8f694948b6a2","Type":"ContainerDied","Data":"ef90acde1c02f78da6ee25dd4a23b978ff700b3a89c4af4bd1a7781af60004f0"} Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.207430 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f459969c7-55hf7" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.259465 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-d5f5b8fd9-mdlh9"] Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.269047 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-d5f5b8fd9-mdlh9"] Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.284421 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-59bc478f75-xdfhn"] Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.289822 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-59bc478f75-xdfhn"] Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.323304 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5f459969c7-55hf7"] Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.330786 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5f459969c7-55hf7"] Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.536598 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-v6lrw"] Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.548833 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-v6lrw"] Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.649034 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-45k2h"] Oct 07 14:03:07 crc kubenswrapper[4959]: E1007 14:03:07.649395 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3205491a-c78b-4eff-b2e1-aa17eb81f5b4" containerName="keystone-bootstrap" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.649409 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3205491a-c78b-4eff-b2e1-aa17eb81f5b4" containerName="keystone-bootstrap" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.649566 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="3205491a-c78b-4eff-b2e1-aa17eb81f5b4" containerName="keystone-bootstrap" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.650120 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-45k2h" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.652379 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.653008 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-z8j6b" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.653506 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.654085 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.661772 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-45k2h"] Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.773046 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hldmq\" (UniqueName: \"kubernetes.io/projected/a359776a-07ed-46e8-b20b-f7b7addaed8d-kube-api-access-hldmq\") pod \"keystone-bootstrap-45k2h\" (UID: \"a359776a-07ed-46e8-b20b-f7b7addaed8d\") " pod="openstack/keystone-bootstrap-45k2h" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.773086 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-scripts\") pod \"keystone-bootstrap-45k2h\" (UID: \"a359776a-07ed-46e8-b20b-f7b7addaed8d\") " pod="openstack/keystone-bootstrap-45k2h" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.773145 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-fernet-keys\") pod \"keystone-bootstrap-45k2h\" (UID: \"a359776a-07ed-46e8-b20b-f7b7addaed8d\") " pod="openstack/keystone-bootstrap-45k2h" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.773207 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-credential-keys\") pod \"keystone-bootstrap-45k2h\" (UID: \"a359776a-07ed-46e8-b20b-f7b7addaed8d\") " pod="openstack/keystone-bootstrap-45k2h" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.773251 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-combined-ca-bundle\") pod \"keystone-bootstrap-45k2h\" (UID: \"a359776a-07ed-46e8-b20b-f7b7addaed8d\") " pod="openstack/keystone-bootstrap-45k2h" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.773976 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-config-data\") pod \"keystone-bootstrap-45k2h\" (UID: \"a359776a-07ed-46e8-b20b-f7b7addaed8d\") " pod="openstack/keystone-bootstrap-45k2h" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.875060 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-config-data\") pod \"keystone-bootstrap-45k2h\" (UID: \"a359776a-07ed-46e8-b20b-f7b7addaed8d\") " pod="openstack/keystone-bootstrap-45k2h" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.875196 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hldmq\" (UniqueName: \"kubernetes.io/projected/a359776a-07ed-46e8-b20b-f7b7addaed8d-kube-api-access-hldmq\") pod \"keystone-bootstrap-45k2h\" (UID: \"a359776a-07ed-46e8-b20b-f7b7addaed8d\") " pod="openstack/keystone-bootstrap-45k2h" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.875226 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-scripts\") pod \"keystone-bootstrap-45k2h\" (UID: \"a359776a-07ed-46e8-b20b-f7b7addaed8d\") " pod="openstack/keystone-bootstrap-45k2h" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.875260 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-fernet-keys\") pod \"keystone-bootstrap-45k2h\" (UID: \"a359776a-07ed-46e8-b20b-f7b7addaed8d\") " pod="openstack/keystone-bootstrap-45k2h" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.875313 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-credential-keys\") pod \"keystone-bootstrap-45k2h\" (UID: \"a359776a-07ed-46e8-b20b-f7b7addaed8d\") " pod="openstack/keystone-bootstrap-45k2h" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.875354 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-combined-ca-bundle\") pod \"keystone-bootstrap-45k2h\" (UID: \"a359776a-07ed-46e8-b20b-f7b7addaed8d\") " pod="openstack/keystone-bootstrap-45k2h" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.881192 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-credential-keys\") pod \"keystone-bootstrap-45k2h\" (UID: \"a359776a-07ed-46e8-b20b-f7b7addaed8d\") " pod="openstack/keystone-bootstrap-45k2h" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.881361 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-config-data\") pod \"keystone-bootstrap-45k2h\" (UID: \"a359776a-07ed-46e8-b20b-f7b7addaed8d\") " pod="openstack/keystone-bootstrap-45k2h" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.881925 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-scripts\") pod \"keystone-bootstrap-45k2h\" (UID: \"a359776a-07ed-46e8-b20b-f7b7addaed8d\") " pod="openstack/keystone-bootstrap-45k2h" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.882224 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-fernet-keys\") pod \"keystone-bootstrap-45k2h\" (UID: \"a359776a-07ed-46e8-b20b-f7b7addaed8d\") " pod="openstack/keystone-bootstrap-45k2h" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.885473 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-combined-ca-bundle\") pod \"keystone-bootstrap-45k2h\" (UID: \"a359776a-07ed-46e8-b20b-f7b7addaed8d\") " pod="openstack/keystone-bootstrap-45k2h" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.905284 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hldmq\" (UniqueName: \"kubernetes.io/projected/a359776a-07ed-46e8-b20b-f7b7addaed8d-kube-api-access-hldmq\") pod \"keystone-bootstrap-45k2h\" (UID: \"a359776a-07ed-46e8-b20b-f7b7addaed8d\") " pod="openstack/keystone-bootstrap-45k2h" Oct 07 14:03:07 crc kubenswrapper[4959]: I1007 14:03:07.968236 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-45k2h" Oct 07 14:03:08 crc kubenswrapper[4959]: I1007 14:03:08.071527 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" Oct 07 14:03:08 crc kubenswrapper[4959]: I1007 14:03:08.178824 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwpvl\" (UniqueName: \"kubernetes.io/projected/4c502b70-d657-47d6-a703-4ecf3cc84d01-kube-api-access-hwpvl\") pod \"4c502b70-d657-47d6-a703-4ecf3cc84d01\" (UID: \"4c502b70-d657-47d6-a703-4ecf3cc84d01\") " Oct 07 14:03:08 crc kubenswrapper[4959]: I1007 14:03:08.178971 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c502b70-d657-47d6-a703-4ecf3cc84d01-ovsdbserver-sb\") pod \"4c502b70-d657-47d6-a703-4ecf3cc84d01\" (UID: \"4c502b70-d657-47d6-a703-4ecf3cc84d01\") " Oct 07 14:03:08 crc kubenswrapper[4959]: I1007 14:03:08.179025 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c502b70-d657-47d6-a703-4ecf3cc84d01-config\") pod \"4c502b70-d657-47d6-a703-4ecf3cc84d01\" (UID: \"4c502b70-d657-47d6-a703-4ecf3cc84d01\") " Oct 07 14:03:08 crc kubenswrapper[4959]: I1007 14:03:08.179155 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c502b70-d657-47d6-a703-4ecf3cc84d01-ovsdbserver-nb\") pod \"4c502b70-d657-47d6-a703-4ecf3cc84d01\" (UID: \"4c502b70-d657-47d6-a703-4ecf3cc84d01\") " Oct 07 14:03:08 crc kubenswrapper[4959]: I1007 14:03:08.179183 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c502b70-d657-47d6-a703-4ecf3cc84d01-dns-svc\") pod \"4c502b70-d657-47d6-a703-4ecf3cc84d01\" (UID: \"4c502b70-d657-47d6-a703-4ecf3cc84d01\") " Oct 07 14:03:08 crc kubenswrapper[4959]: I1007 14:03:08.182539 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c502b70-d657-47d6-a703-4ecf3cc84d01-kube-api-access-hwpvl" (OuterVolumeSpecName: "kube-api-access-hwpvl") pod "4c502b70-d657-47d6-a703-4ecf3cc84d01" (UID: "4c502b70-d657-47d6-a703-4ecf3cc84d01"). InnerVolumeSpecName "kube-api-access-hwpvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:03:08 crc kubenswrapper[4959]: E1007 14:03:08.214399 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:85c75d60e1bd2f8a9ea0a2bb21a8df64c0a6f7b504cc1a05a355981d4b90e92f" Oct 07 14:03:08 crc kubenswrapper[4959]: E1007 14:03:08.214884 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:85c75d60e1bd2f8a9ea0a2bb21a8df64c0a6f7b504cc1a05a355981d4b90e92f,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zmzf9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-jfrdg_openstack(7b70a321-ffb7-429f-b825-4cd872ded9ff): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 14:03:08 crc kubenswrapper[4959]: E1007 14:03:08.215977 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-jfrdg" podUID="7b70a321-ffb7-429f-b825-4cd872ded9ff" Oct 07 14:03:08 crc kubenswrapper[4959]: I1007 14:03:08.220251 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" event={"ID":"4c502b70-d657-47d6-a703-4ecf3cc84d01","Type":"ContainerDied","Data":"742deb80e08dc504344a63df2479689c408d2162fbedec5d05274f611d49dd3e"} Oct 07 14:03:08 crc kubenswrapper[4959]: I1007 14:03:08.220307 4959 scope.go:117] "RemoveContainer" containerID="a8a2200da57e5e47c36da78d602f44fd8a60bcac7cad79714ee6b62fcf3956d8" Oct 07 14:03:08 crc kubenswrapper[4959]: I1007 14:03:08.220452 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc45f6dcf-t6p7q" Oct 07 14:03:08 crc kubenswrapper[4959]: I1007 14:03:08.225287 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c502b70-d657-47d6-a703-4ecf3cc84d01-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4c502b70-d657-47d6-a703-4ecf3cc84d01" (UID: "4c502b70-d657-47d6-a703-4ecf3cc84d01"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:03:08 crc kubenswrapper[4959]: I1007 14:03:08.226338 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c502b70-d657-47d6-a703-4ecf3cc84d01-config" (OuterVolumeSpecName: "config") pod "4c502b70-d657-47d6-a703-4ecf3cc84d01" (UID: "4c502b70-d657-47d6-a703-4ecf3cc84d01"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:03:08 crc kubenswrapper[4959]: I1007 14:03:08.227293 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c502b70-d657-47d6-a703-4ecf3cc84d01-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4c502b70-d657-47d6-a703-4ecf3cc84d01" (UID: "4c502b70-d657-47d6-a703-4ecf3cc84d01"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:03:08 crc kubenswrapper[4959]: I1007 14:03:08.228173 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c502b70-d657-47d6-a703-4ecf3cc84d01-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4c502b70-d657-47d6-a703-4ecf3cc84d01" (UID: "4c502b70-d657-47d6-a703-4ecf3cc84d01"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:03:08 crc kubenswrapper[4959]: I1007 14:03:08.282841 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c502b70-d657-47d6-a703-4ecf3cc84d01-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:08 crc kubenswrapper[4959]: I1007 14:03:08.282881 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c502b70-d657-47d6-a703-4ecf3cc84d01-config\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:08 crc kubenswrapper[4959]: I1007 14:03:08.282895 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c502b70-d657-47d6-a703-4ecf3cc84d01-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:08 crc kubenswrapper[4959]: I1007 14:03:08.282909 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c502b70-d657-47d6-a703-4ecf3cc84d01-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:08 crc kubenswrapper[4959]: I1007 14:03:08.282922 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwpvl\" (UniqueName: \"kubernetes.io/projected/4c502b70-d657-47d6-a703-4ecf3cc84d01-kube-api-access-hwpvl\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:08 crc kubenswrapper[4959]: I1007 14:03:08.560778 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bc45f6dcf-t6p7q"] Oct 07 14:03:08 crc kubenswrapper[4959]: I1007 14:03:08.567576 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bc45f6dcf-t6p7q"] Oct 07 14:03:08 crc kubenswrapper[4959]: I1007 14:03:08.664674 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3205491a-c78b-4eff-b2e1-aa17eb81f5b4" path="/var/lib/kubelet/pods/3205491a-c78b-4eff-b2e1-aa17eb81f5b4/volumes" Oct 07 14:03:08 crc kubenswrapper[4959]: I1007 14:03:08.665595 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c502b70-d657-47d6-a703-4ecf3cc84d01" path="/var/lib/kubelet/pods/4c502b70-d657-47d6-a703-4ecf3cc84d01/volumes" Oct 07 14:03:08 crc kubenswrapper[4959]: I1007 14:03:08.667277 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="594da4a6-8781-416f-abdc-8f694948b6a2" path="/var/lib/kubelet/pods/594da4a6-8781-416f-abdc-8f694948b6a2/volumes" Oct 07 14:03:08 crc kubenswrapper[4959]: I1007 14:03:08.668400 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0" path="/var/lib/kubelet/pods/5ca2b859-cf47-4b29-8de5-d2fbbaad5bb0/volumes" Oct 07 14:03:08 crc kubenswrapper[4959]: I1007 14:03:08.668953 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="906180a5-7c20-4109-8094-7d578d8ad374" path="/var/lib/kubelet/pods/906180a5-7c20-4109-8094-7d578d8ad374/volumes" Oct 07 14:03:08 crc kubenswrapper[4959]: I1007 14:03:08.718636 4959 scope.go:117] "RemoveContainer" containerID="60b68bdc6665fe51a22286c453a65af6ffea0a6c2901f698c050116f76fbc189" Oct 07 14:03:09 crc kubenswrapper[4959]: I1007 14:03:09.137757 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-744bf8f46d-6twnn"] Oct 07 14:03:09 crc kubenswrapper[4959]: I1007 14:03:09.160376 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-76d6fc6cbd-2g7sq"] Oct 07 14:03:09 crc kubenswrapper[4959]: I1007 14:03:09.232229 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ff2nz" event={"ID":"2fcce932-e96f-4a33-9511-d84540bd98fa","Type":"ContainerStarted","Data":"66798f105e40cb45703200b22d95be4846111976ce149d676b6c3f12a94de0f4"} Oct 07 14:03:09 crc kubenswrapper[4959]: I1007 14:03:09.234838 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-744bf8f46d-6twnn" event={"ID":"a588fbdc-fd65-49f6-ab59-4901c7ca690a","Type":"ContainerStarted","Data":"0e8c421e3ea812af0ca6b26318aada327d562316382b23587cf36edf6e88d253"} Oct 07 14:03:09 crc kubenswrapper[4959]: I1007 14:03:09.236035 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76d6fc6cbd-2g7sq" event={"ID":"13b27eed-f50f-4474-ace5-0f12e733f6cf","Type":"ContainerStarted","Data":"8aee9515e15b3967ca7f74738796ab0be31a45363bc766612da3b33f2d2c2934"} Oct 07 14:03:09 crc kubenswrapper[4959]: E1007 14:03:09.238320 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:85c75d60e1bd2f8a9ea0a2bb21a8df64c0a6f7b504cc1a05a355981d4b90e92f\\\"\"" pod="openstack/cinder-db-sync-jfrdg" podUID="7b70a321-ffb7-429f-b825-4cd872ded9ff" Oct 07 14:03:09 crc kubenswrapper[4959]: I1007 14:03:09.294704 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-45k2h"] Oct 07 14:03:10 crc kubenswrapper[4959]: I1007 14:03:10.243617 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-45k2h" event={"ID":"a359776a-07ed-46e8-b20b-f7b7addaed8d","Type":"ContainerStarted","Data":"3b6b9b26c429a499cf7ec26d7c0fa5d05bba2f7a179c871fd6081cff7c6ffe21"} Oct 07 14:03:10 crc kubenswrapper[4959]: I1007 14:03:10.264533 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-ff2nz" podStartSLOduration=3.663755911 podStartE2EDuration="37.264516014s" podCreationTimestamp="2025-10-07 14:02:33 +0000 UTC" firstStartedPulling="2025-10-07 14:02:34.377954968 +0000 UTC m=+1036.461359293" lastFinishedPulling="2025-10-07 14:03:07.978715071 +0000 UTC m=+1070.062119396" observedRunningTime="2025-10-07 14:03:10.257801418 +0000 UTC m=+1072.341205783" watchObservedRunningTime="2025-10-07 14:03:10.264516014 +0000 UTC m=+1072.347920339" Oct 07 14:03:11 crc kubenswrapper[4959]: I1007 14:03:11.253854 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-45k2h" event={"ID":"a359776a-07ed-46e8-b20b-f7b7addaed8d","Type":"ContainerStarted","Data":"6361a07ae0f69ab1a40e597bf3eb451fd4fbeedc2ddf0353d49a648fca005ca9"} Oct 07 14:03:12 crc kubenswrapper[4959]: I1007 14:03:12.263264 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd33b52c-3f7f-425e-aa42-7342343067a6","Type":"ContainerStarted","Data":"a7768743c10e8cceb2e9f59d7ad327c68847c768df2193deb859955e0f76a5de"} Oct 07 14:03:12 crc kubenswrapper[4959]: I1007 14:03:12.292696 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-45k2h" podStartSLOduration=5.292666507 podStartE2EDuration="5.292666507s" podCreationTimestamp="2025-10-07 14:03:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:03:12.284658989 +0000 UTC m=+1074.368063354" watchObservedRunningTime="2025-10-07 14:03:12.292666507 +0000 UTC m=+1074.376070882" Oct 07 14:03:27 crc kubenswrapper[4959]: I1007 14:03:27.417565 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd33b52c-3f7f-425e-aa42-7342343067a6","Type":"ContainerStarted","Data":"1395f86ab5b523c157ccdd52e29a4d040ff47bbc2a8f66c3696b941b2aa1e1eb"} Oct 07 14:03:27 crc kubenswrapper[4959]: I1007 14:03:27.418640 4959 generic.go:334] "Generic (PLEG): container finished" podID="a359776a-07ed-46e8-b20b-f7b7addaed8d" containerID="6361a07ae0f69ab1a40e597bf3eb451fd4fbeedc2ddf0353d49a648fca005ca9" exitCode=0 Oct 07 14:03:27 crc kubenswrapper[4959]: I1007 14:03:27.418703 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-45k2h" event={"ID":"a359776a-07ed-46e8-b20b-f7b7addaed8d","Type":"ContainerDied","Data":"6361a07ae0f69ab1a40e597bf3eb451fd4fbeedc2ddf0353d49a648fca005ca9"} Oct 07 14:03:27 crc kubenswrapper[4959]: I1007 14:03:27.420059 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-744bf8f46d-6twnn" event={"ID":"a588fbdc-fd65-49f6-ab59-4901c7ca690a","Type":"ContainerStarted","Data":"76d664db4d06017fdcd093f711c6c672a5feacc3d610e94a71a01f0793720247"} Oct 07 14:03:27 crc kubenswrapper[4959]: I1007 14:03:27.420117 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-744bf8f46d-6twnn" event={"ID":"a588fbdc-fd65-49f6-ab59-4901c7ca690a","Type":"ContainerStarted","Data":"0b8d35aaa990504cd841339a7496639b4a5b31cf2af6d3b62946d1c9d29fa02f"} Oct 07 14:03:27 crc kubenswrapper[4959]: I1007 14:03:27.421741 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76d6fc6cbd-2g7sq" event={"ID":"13b27eed-f50f-4474-ace5-0f12e733f6cf","Type":"ContainerStarted","Data":"09e80647ebeefb52e1f83d12aa00630658e7d866f72883ade6e95d6c5dca928a"} Oct 07 14:03:27 crc kubenswrapper[4959]: I1007 14:03:27.421776 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76d6fc6cbd-2g7sq" event={"ID":"13b27eed-f50f-4474-ace5-0f12e733f6cf","Type":"ContainerStarted","Data":"0548f45657b38e3cea35daa5e8fc0cf810a21da91f7bd242dc45a8176c167f5a"} Oct 07 14:03:27 crc kubenswrapper[4959]: I1007 14:03:27.423879 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2hhl9" event={"ID":"c3b8b961-3c0e-47dc-b28f-d55c9821ab34","Type":"ContainerStarted","Data":"88d73562583ddda94182356c56de7337fd18fc33948a51321d001c62cd5d8ebc"} Oct 07 14:03:27 crc kubenswrapper[4959]: I1007 14:03:27.425512 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-jfrdg" event={"ID":"7b70a321-ffb7-429f-b825-4cd872ded9ff","Type":"ContainerStarted","Data":"1c98cd878dfdc18d00933f385480ff055b149f850fe48ee3960f16c3ce508cb7"} Oct 07 14:03:27 crc kubenswrapper[4959]: I1007 14:03:27.461344 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-76d6fc6cbd-2g7sq" podStartSLOduration=24.019877207 podStartE2EDuration="41.461324751s" podCreationTimestamp="2025-10-07 14:02:46 +0000 UTC" firstStartedPulling="2025-10-07 14:03:09.146705979 +0000 UTC m=+1071.230110314" lastFinishedPulling="2025-10-07 14:03:26.588153533 +0000 UTC m=+1088.671557858" observedRunningTime="2025-10-07 14:03:27.459610489 +0000 UTC m=+1089.543014824" watchObservedRunningTime="2025-10-07 14:03:27.461324751 +0000 UTC m=+1089.544729076" Oct 07 14:03:27 crc kubenswrapper[4959]: I1007 14:03:27.483457 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-744bf8f46d-6twnn" podStartSLOduration=27.568388539 podStartE2EDuration="41.483438169s" podCreationTimestamp="2025-10-07 14:02:46 +0000 UTC" firstStartedPulling="2025-10-07 14:03:09.138612938 +0000 UTC m=+1071.222017273" lastFinishedPulling="2025-10-07 14:03:23.053662528 +0000 UTC m=+1085.137066903" observedRunningTime="2025-10-07 14:03:27.479514702 +0000 UTC m=+1089.562919027" watchObservedRunningTime="2025-10-07 14:03:27.483438169 +0000 UTC m=+1089.566842494" Oct 07 14:03:27 crc kubenswrapper[4959]: I1007 14:03:27.496442 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-2hhl9" podStartSLOduration=3.164023052 podStartE2EDuration="52.49642423s" podCreationTimestamp="2025-10-07 14:02:35 +0000 UTC" firstStartedPulling="2025-10-07 14:02:36.746317262 +0000 UTC m=+1038.829721587" lastFinishedPulling="2025-10-07 14:03:26.07871844 +0000 UTC m=+1088.162122765" observedRunningTime="2025-10-07 14:03:27.494276287 +0000 UTC m=+1089.577680612" watchObservedRunningTime="2025-10-07 14:03:27.49642423 +0000 UTC m=+1089.579828555" Oct 07 14:03:27 crc kubenswrapper[4959]: I1007 14:03:27.514384 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-jfrdg" podStartSLOduration=2.737711905 podStartE2EDuration="52.514366115s" podCreationTimestamp="2025-10-07 14:02:35 +0000 UTC" firstStartedPulling="2025-10-07 14:02:36.838356588 +0000 UTC m=+1038.921760913" lastFinishedPulling="2025-10-07 14:03:26.615010788 +0000 UTC m=+1088.698415123" observedRunningTime="2025-10-07 14:03:27.507889754 +0000 UTC m=+1089.591294079" watchObservedRunningTime="2025-10-07 14:03:27.514366115 +0000 UTC m=+1089.597770440" Oct 07 14:03:28 crc kubenswrapper[4959]: I1007 14:03:28.439249 4959 generic.go:334] "Generic (PLEG): container finished" podID="2fcce932-e96f-4a33-9511-d84540bd98fa" containerID="66798f105e40cb45703200b22d95be4846111976ce149d676b6c3f12a94de0f4" exitCode=0 Oct 07 14:03:28 crc kubenswrapper[4959]: I1007 14:03:28.439399 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ff2nz" event={"ID":"2fcce932-e96f-4a33-9511-d84540bd98fa","Type":"ContainerDied","Data":"66798f105e40cb45703200b22d95be4846111976ce149d676b6c3f12a94de0f4"} Oct 07 14:03:28 crc kubenswrapper[4959]: I1007 14:03:28.795092 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-45k2h" Oct 07 14:03:28 crc kubenswrapper[4959]: I1007 14:03:28.983263 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-config-data\") pod \"a359776a-07ed-46e8-b20b-f7b7addaed8d\" (UID: \"a359776a-07ed-46e8-b20b-f7b7addaed8d\") " Oct 07 14:03:28 crc kubenswrapper[4959]: I1007 14:03:28.983373 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-combined-ca-bundle\") pod \"a359776a-07ed-46e8-b20b-f7b7addaed8d\" (UID: \"a359776a-07ed-46e8-b20b-f7b7addaed8d\") " Oct 07 14:03:28 crc kubenswrapper[4959]: I1007 14:03:28.983475 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-scripts\") pod \"a359776a-07ed-46e8-b20b-f7b7addaed8d\" (UID: \"a359776a-07ed-46e8-b20b-f7b7addaed8d\") " Oct 07 14:03:28 crc kubenswrapper[4959]: I1007 14:03:28.983500 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hldmq\" (UniqueName: \"kubernetes.io/projected/a359776a-07ed-46e8-b20b-f7b7addaed8d-kube-api-access-hldmq\") pod \"a359776a-07ed-46e8-b20b-f7b7addaed8d\" (UID: \"a359776a-07ed-46e8-b20b-f7b7addaed8d\") " Oct 07 14:03:28 crc kubenswrapper[4959]: I1007 14:03:28.983518 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-credential-keys\") pod \"a359776a-07ed-46e8-b20b-f7b7addaed8d\" (UID: \"a359776a-07ed-46e8-b20b-f7b7addaed8d\") " Oct 07 14:03:28 crc kubenswrapper[4959]: I1007 14:03:28.983534 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-fernet-keys\") pod \"a359776a-07ed-46e8-b20b-f7b7addaed8d\" (UID: \"a359776a-07ed-46e8-b20b-f7b7addaed8d\") " Oct 07 14:03:28 crc kubenswrapper[4959]: I1007 14:03:28.989429 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-scripts" (OuterVolumeSpecName: "scripts") pod "a359776a-07ed-46e8-b20b-f7b7addaed8d" (UID: "a359776a-07ed-46e8-b20b-f7b7addaed8d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:28 crc kubenswrapper[4959]: I1007 14:03:28.989959 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a359776a-07ed-46e8-b20b-f7b7addaed8d-kube-api-access-hldmq" (OuterVolumeSpecName: "kube-api-access-hldmq") pod "a359776a-07ed-46e8-b20b-f7b7addaed8d" (UID: "a359776a-07ed-46e8-b20b-f7b7addaed8d"). InnerVolumeSpecName "kube-api-access-hldmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:03:28 crc kubenswrapper[4959]: I1007 14:03:28.989954 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a359776a-07ed-46e8-b20b-f7b7addaed8d" (UID: "a359776a-07ed-46e8-b20b-f7b7addaed8d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:28 crc kubenswrapper[4959]: I1007 14:03:28.994218 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "a359776a-07ed-46e8-b20b-f7b7addaed8d" (UID: "a359776a-07ed-46e8-b20b-f7b7addaed8d"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.008394 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a359776a-07ed-46e8-b20b-f7b7addaed8d" (UID: "a359776a-07ed-46e8-b20b-f7b7addaed8d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.017222 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-config-data" (OuterVolumeSpecName: "config-data") pod "a359776a-07ed-46e8-b20b-f7b7addaed8d" (UID: "a359776a-07ed-46e8-b20b-f7b7addaed8d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.085314 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.085345 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hldmq\" (UniqueName: \"kubernetes.io/projected/a359776a-07ed-46e8-b20b-f7b7addaed8d-kube-api-access-hldmq\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.085358 4959 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.085367 4959 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.085376 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.085384 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a359776a-07ed-46e8-b20b-f7b7addaed8d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.467168 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-45k2h" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.467311 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-45k2h" event={"ID":"a359776a-07ed-46e8-b20b-f7b7addaed8d","Type":"ContainerDied","Data":"3b6b9b26c429a499cf7ec26d7c0fa5d05bba2f7a179c871fd6081cff7c6ffe21"} Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.467366 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b6b9b26c429a499cf7ec26d7c0fa5d05bba2f7a179c871fd6081cff7c6ffe21" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.549440 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5c77f969b5-4w2xs"] Oct 07 14:03:29 crc kubenswrapper[4959]: E1007 14:03:29.549774 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c502b70-d657-47d6-a703-4ecf3cc84d01" containerName="init" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.549790 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c502b70-d657-47d6-a703-4ecf3cc84d01" containerName="init" Oct 07 14:03:29 crc kubenswrapper[4959]: E1007 14:03:29.549802 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c502b70-d657-47d6-a703-4ecf3cc84d01" containerName="dnsmasq-dns" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.549809 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c502b70-d657-47d6-a703-4ecf3cc84d01" containerName="dnsmasq-dns" Oct 07 14:03:29 crc kubenswrapper[4959]: E1007 14:03:29.549827 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a359776a-07ed-46e8-b20b-f7b7addaed8d" containerName="keystone-bootstrap" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.549834 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a359776a-07ed-46e8-b20b-f7b7addaed8d" containerName="keystone-bootstrap" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.549987 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c502b70-d657-47d6-a703-4ecf3cc84d01" containerName="dnsmasq-dns" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.550014 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a359776a-07ed-46e8-b20b-f7b7addaed8d" containerName="keystone-bootstrap" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.550527 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5c77f969b5-4w2xs" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.552625 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.552761 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.553870 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5c77f969b5-4w2xs"] Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.554069 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.554381 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.554491 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.554638 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-z8j6b" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.598763 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33e694c5-266c-40e1-a805-c174cd094645-config-data\") pod \"keystone-5c77f969b5-4w2xs\" (UID: \"33e694c5-266c-40e1-a805-c174cd094645\") " pod="openstack/keystone-5c77f969b5-4w2xs" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.598804 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33e694c5-266c-40e1-a805-c174cd094645-scripts\") pod \"keystone-5c77f969b5-4w2xs\" (UID: \"33e694c5-266c-40e1-a805-c174cd094645\") " pod="openstack/keystone-5c77f969b5-4w2xs" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.598826 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33e694c5-266c-40e1-a805-c174cd094645-combined-ca-bundle\") pod \"keystone-5c77f969b5-4w2xs\" (UID: \"33e694c5-266c-40e1-a805-c174cd094645\") " pod="openstack/keystone-5c77f969b5-4w2xs" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.598922 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n5tz\" (UniqueName: \"kubernetes.io/projected/33e694c5-266c-40e1-a805-c174cd094645-kube-api-access-5n5tz\") pod \"keystone-5c77f969b5-4w2xs\" (UID: \"33e694c5-266c-40e1-a805-c174cd094645\") " pod="openstack/keystone-5c77f969b5-4w2xs" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.598952 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/33e694c5-266c-40e1-a805-c174cd094645-credential-keys\") pod \"keystone-5c77f969b5-4w2xs\" (UID: \"33e694c5-266c-40e1-a805-c174cd094645\") " pod="openstack/keystone-5c77f969b5-4w2xs" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.598992 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/33e694c5-266c-40e1-a805-c174cd094645-public-tls-certs\") pod \"keystone-5c77f969b5-4w2xs\" (UID: \"33e694c5-266c-40e1-a805-c174cd094645\") " pod="openstack/keystone-5c77f969b5-4w2xs" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.599038 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/33e694c5-266c-40e1-a805-c174cd094645-fernet-keys\") pod \"keystone-5c77f969b5-4w2xs\" (UID: \"33e694c5-266c-40e1-a805-c174cd094645\") " pod="openstack/keystone-5c77f969b5-4w2xs" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.599143 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33e694c5-266c-40e1-a805-c174cd094645-internal-tls-certs\") pod \"keystone-5c77f969b5-4w2xs\" (UID: \"33e694c5-266c-40e1-a805-c174cd094645\") " pod="openstack/keystone-5c77f969b5-4w2xs" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.700956 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33e694c5-266c-40e1-a805-c174cd094645-config-data\") pod \"keystone-5c77f969b5-4w2xs\" (UID: \"33e694c5-266c-40e1-a805-c174cd094645\") " pod="openstack/keystone-5c77f969b5-4w2xs" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.701002 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33e694c5-266c-40e1-a805-c174cd094645-scripts\") pod \"keystone-5c77f969b5-4w2xs\" (UID: \"33e694c5-266c-40e1-a805-c174cd094645\") " pod="openstack/keystone-5c77f969b5-4w2xs" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.701024 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33e694c5-266c-40e1-a805-c174cd094645-combined-ca-bundle\") pod \"keystone-5c77f969b5-4w2xs\" (UID: \"33e694c5-266c-40e1-a805-c174cd094645\") " pod="openstack/keystone-5c77f969b5-4w2xs" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.701071 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n5tz\" (UniqueName: \"kubernetes.io/projected/33e694c5-266c-40e1-a805-c174cd094645-kube-api-access-5n5tz\") pod \"keystone-5c77f969b5-4w2xs\" (UID: \"33e694c5-266c-40e1-a805-c174cd094645\") " pod="openstack/keystone-5c77f969b5-4w2xs" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.701087 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/33e694c5-266c-40e1-a805-c174cd094645-credential-keys\") pod \"keystone-5c77f969b5-4w2xs\" (UID: \"33e694c5-266c-40e1-a805-c174cd094645\") " pod="openstack/keystone-5c77f969b5-4w2xs" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.701130 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/33e694c5-266c-40e1-a805-c174cd094645-public-tls-certs\") pod \"keystone-5c77f969b5-4w2xs\" (UID: \"33e694c5-266c-40e1-a805-c174cd094645\") " pod="openstack/keystone-5c77f969b5-4w2xs" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.701155 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/33e694c5-266c-40e1-a805-c174cd094645-fernet-keys\") pod \"keystone-5c77f969b5-4w2xs\" (UID: \"33e694c5-266c-40e1-a805-c174cd094645\") " pod="openstack/keystone-5c77f969b5-4w2xs" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.703007 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33e694c5-266c-40e1-a805-c174cd094645-internal-tls-certs\") pod \"keystone-5c77f969b5-4w2xs\" (UID: \"33e694c5-266c-40e1-a805-c174cd094645\") " pod="openstack/keystone-5c77f969b5-4w2xs" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.719153 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/33e694c5-266c-40e1-a805-c174cd094645-credential-keys\") pod \"keystone-5c77f969b5-4w2xs\" (UID: \"33e694c5-266c-40e1-a805-c174cd094645\") " pod="openstack/keystone-5c77f969b5-4w2xs" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.719669 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33e694c5-266c-40e1-a805-c174cd094645-config-data\") pod \"keystone-5c77f969b5-4w2xs\" (UID: \"33e694c5-266c-40e1-a805-c174cd094645\") " pod="openstack/keystone-5c77f969b5-4w2xs" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.744912 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33e694c5-266c-40e1-a805-c174cd094645-internal-tls-certs\") pod \"keystone-5c77f969b5-4w2xs\" (UID: \"33e694c5-266c-40e1-a805-c174cd094645\") " pod="openstack/keystone-5c77f969b5-4w2xs" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.746265 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33e694c5-266c-40e1-a805-c174cd094645-scripts\") pod \"keystone-5c77f969b5-4w2xs\" (UID: \"33e694c5-266c-40e1-a805-c174cd094645\") " pod="openstack/keystone-5c77f969b5-4w2xs" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.746835 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/33e694c5-266c-40e1-a805-c174cd094645-public-tls-certs\") pod \"keystone-5c77f969b5-4w2xs\" (UID: \"33e694c5-266c-40e1-a805-c174cd094645\") " pod="openstack/keystone-5c77f969b5-4w2xs" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.749763 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33e694c5-266c-40e1-a805-c174cd094645-combined-ca-bundle\") pod \"keystone-5c77f969b5-4w2xs\" (UID: \"33e694c5-266c-40e1-a805-c174cd094645\") " pod="openstack/keystone-5c77f969b5-4w2xs" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.750289 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/33e694c5-266c-40e1-a805-c174cd094645-fernet-keys\") pod \"keystone-5c77f969b5-4w2xs\" (UID: \"33e694c5-266c-40e1-a805-c174cd094645\") " pod="openstack/keystone-5c77f969b5-4w2xs" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.756714 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n5tz\" (UniqueName: \"kubernetes.io/projected/33e694c5-266c-40e1-a805-c174cd094645-kube-api-access-5n5tz\") pod \"keystone-5c77f969b5-4w2xs\" (UID: \"33e694c5-266c-40e1-a805-c174cd094645\") " pod="openstack/keystone-5c77f969b5-4w2xs" Oct 07 14:03:29 crc kubenswrapper[4959]: I1007 14:03:29.916599 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5c77f969b5-4w2xs" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.035655 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ff2nz" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.211608 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fcce932-e96f-4a33-9511-d84540bd98fa-logs\") pod \"2fcce932-e96f-4a33-9511-d84540bd98fa\" (UID: \"2fcce932-e96f-4a33-9511-d84540bd98fa\") " Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.211705 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fcce932-e96f-4a33-9511-d84540bd98fa-config-data\") pod \"2fcce932-e96f-4a33-9511-d84540bd98fa\" (UID: \"2fcce932-e96f-4a33-9511-d84540bd98fa\") " Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.211768 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fcce932-e96f-4a33-9511-d84540bd98fa-scripts\") pod \"2fcce932-e96f-4a33-9511-d84540bd98fa\" (UID: \"2fcce932-e96f-4a33-9511-d84540bd98fa\") " Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.211877 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqwlb\" (UniqueName: \"kubernetes.io/projected/2fcce932-e96f-4a33-9511-d84540bd98fa-kube-api-access-dqwlb\") pod \"2fcce932-e96f-4a33-9511-d84540bd98fa\" (UID: \"2fcce932-e96f-4a33-9511-d84540bd98fa\") " Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.211922 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fcce932-e96f-4a33-9511-d84540bd98fa-combined-ca-bundle\") pod \"2fcce932-e96f-4a33-9511-d84540bd98fa\" (UID: \"2fcce932-e96f-4a33-9511-d84540bd98fa\") " Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.212616 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fcce932-e96f-4a33-9511-d84540bd98fa-logs" (OuterVolumeSpecName: "logs") pod "2fcce932-e96f-4a33-9511-d84540bd98fa" (UID: "2fcce932-e96f-4a33-9511-d84540bd98fa"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.218496 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fcce932-e96f-4a33-9511-d84540bd98fa-scripts" (OuterVolumeSpecName: "scripts") pod "2fcce932-e96f-4a33-9511-d84540bd98fa" (UID: "2fcce932-e96f-4a33-9511-d84540bd98fa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.222007 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fcce932-e96f-4a33-9511-d84540bd98fa-kube-api-access-dqwlb" (OuterVolumeSpecName: "kube-api-access-dqwlb") pod "2fcce932-e96f-4a33-9511-d84540bd98fa" (UID: "2fcce932-e96f-4a33-9511-d84540bd98fa"). InnerVolumeSpecName "kube-api-access-dqwlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.238847 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fcce932-e96f-4a33-9511-d84540bd98fa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2fcce932-e96f-4a33-9511-d84540bd98fa" (UID: "2fcce932-e96f-4a33-9511-d84540bd98fa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.245668 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fcce932-e96f-4a33-9511-d84540bd98fa-config-data" (OuterVolumeSpecName: "config-data") pod "2fcce932-e96f-4a33-9511-d84540bd98fa" (UID: "2fcce932-e96f-4a33-9511-d84540bd98fa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.313327 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fcce932-e96f-4a33-9511-d84540bd98fa-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.313364 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fcce932-e96f-4a33-9511-d84540bd98fa-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.313375 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqwlb\" (UniqueName: \"kubernetes.io/projected/2fcce932-e96f-4a33-9511-d84540bd98fa-kube-api-access-dqwlb\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.313386 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fcce932-e96f-4a33-9511-d84540bd98fa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.313396 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fcce932-e96f-4a33-9511-d84540bd98fa-logs\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.356576 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5c77f969b5-4w2xs"] Oct 07 14:03:30 crc kubenswrapper[4959]: W1007 14:03:30.359588 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33e694c5_266c_40e1_a805_c174cd094645.slice/crio-edae513bbc3866a5a4c519531f7b0c2fcc4e4a8985e2354484ad22855836e3ae WatchSource:0}: Error finding container edae513bbc3866a5a4c519531f7b0c2fcc4e4a8985e2354484ad22855836e3ae: Status 404 returned error can't find the container with id edae513bbc3866a5a4c519531f7b0c2fcc4e4a8985e2354484ad22855836e3ae Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.476774 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5c77f969b5-4w2xs" event={"ID":"33e694c5-266c-40e1-a805-c174cd094645","Type":"ContainerStarted","Data":"edae513bbc3866a5a4c519531f7b0c2fcc4e4a8985e2354484ad22855836e3ae"} Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.479520 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ff2nz" event={"ID":"2fcce932-e96f-4a33-9511-d84540bd98fa","Type":"ContainerDied","Data":"c5099f9b7273b37cc0577cb3ab3f9a6765d4643ebfca51d912543792e8040529"} Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.479553 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5099f9b7273b37cc0577cb3ab3f9a6765d4643ebfca51d912543792e8040529" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.479603 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ff2nz" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.481350 4959 generic.go:334] "Generic (PLEG): container finished" podID="c3b8b961-3c0e-47dc-b28f-d55c9821ab34" containerID="88d73562583ddda94182356c56de7337fd18fc33948a51321d001c62cd5d8ebc" exitCode=0 Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.481389 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2hhl9" event={"ID":"c3b8b961-3c0e-47dc-b28f-d55c9821ab34","Type":"ContainerDied","Data":"88d73562583ddda94182356c56de7337fd18fc33948a51321d001c62cd5d8ebc"} Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.634791 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5c4946988b-h259p"] Oct 07 14:03:30 crc kubenswrapper[4959]: E1007 14:03:30.635321 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fcce932-e96f-4a33-9511-d84540bd98fa" containerName="placement-db-sync" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.635348 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fcce932-e96f-4a33-9511-d84540bd98fa" containerName="placement-db-sync" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.635581 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fcce932-e96f-4a33-9511-d84540bd98fa" containerName="placement-db-sync" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.636724 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5c4946988b-h259p" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.639260 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.640140 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.640319 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.641375 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.642177 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-lv2f8" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.648373 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5c4946988b-h259p"] Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.823440 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvvgp\" (UniqueName: \"kubernetes.io/projected/342978a1-dc32-4347-bf34-6782c201f033-kube-api-access-zvvgp\") pod \"placement-5c4946988b-h259p\" (UID: \"342978a1-dc32-4347-bf34-6782c201f033\") " pod="openstack/placement-5c4946988b-h259p" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.823486 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/342978a1-dc32-4347-bf34-6782c201f033-config-data\") pod \"placement-5c4946988b-h259p\" (UID: \"342978a1-dc32-4347-bf34-6782c201f033\") " pod="openstack/placement-5c4946988b-h259p" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.823525 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/342978a1-dc32-4347-bf34-6782c201f033-combined-ca-bundle\") pod \"placement-5c4946988b-h259p\" (UID: \"342978a1-dc32-4347-bf34-6782c201f033\") " pod="openstack/placement-5c4946988b-h259p" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.823597 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/342978a1-dc32-4347-bf34-6782c201f033-logs\") pod \"placement-5c4946988b-h259p\" (UID: \"342978a1-dc32-4347-bf34-6782c201f033\") " pod="openstack/placement-5c4946988b-h259p" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.823648 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/342978a1-dc32-4347-bf34-6782c201f033-scripts\") pod \"placement-5c4946988b-h259p\" (UID: \"342978a1-dc32-4347-bf34-6782c201f033\") " pod="openstack/placement-5c4946988b-h259p" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.823774 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/342978a1-dc32-4347-bf34-6782c201f033-public-tls-certs\") pod \"placement-5c4946988b-h259p\" (UID: \"342978a1-dc32-4347-bf34-6782c201f033\") " pod="openstack/placement-5c4946988b-h259p" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.823810 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/342978a1-dc32-4347-bf34-6782c201f033-internal-tls-certs\") pod \"placement-5c4946988b-h259p\" (UID: \"342978a1-dc32-4347-bf34-6782c201f033\") " pod="openstack/placement-5c4946988b-h259p" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.925332 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/342978a1-dc32-4347-bf34-6782c201f033-internal-tls-certs\") pod \"placement-5c4946988b-h259p\" (UID: \"342978a1-dc32-4347-bf34-6782c201f033\") " pod="openstack/placement-5c4946988b-h259p" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.925444 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvvgp\" (UniqueName: \"kubernetes.io/projected/342978a1-dc32-4347-bf34-6782c201f033-kube-api-access-zvvgp\") pod \"placement-5c4946988b-h259p\" (UID: \"342978a1-dc32-4347-bf34-6782c201f033\") " pod="openstack/placement-5c4946988b-h259p" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.925466 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/342978a1-dc32-4347-bf34-6782c201f033-config-data\") pod \"placement-5c4946988b-h259p\" (UID: \"342978a1-dc32-4347-bf34-6782c201f033\") " pod="openstack/placement-5c4946988b-h259p" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.925818 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/342978a1-dc32-4347-bf34-6782c201f033-combined-ca-bundle\") pod \"placement-5c4946988b-h259p\" (UID: \"342978a1-dc32-4347-bf34-6782c201f033\") " pod="openstack/placement-5c4946988b-h259p" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.926392 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/342978a1-dc32-4347-bf34-6782c201f033-logs\") pod \"placement-5c4946988b-h259p\" (UID: \"342978a1-dc32-4347-bf34-6782c201f033\") " pod="openstack/placement-5c4946988b-h259p" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.926421 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/342978a1-dc32-4347-bf34-6782c201f033-scripts\") pod \"placement-5c4946988b-h259p\" (UID: \"342978a1-dc32-4347-bf34-6782c201f033\") " pod="openstack/placement-5c4946988b-h259p" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.926448 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/342978a1-dc32-4347-bf34-6782c201f033-public-tls-certs\") pod \"placement-5c4946988b-h259p\" (UID: \"342978a1-dc32-4347-bf34-6782c201f033\") " pod="openstack/placement-5c4946988b-h259p" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.926703 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/342978a1-dc32-4347-bf34-6782c201f033-logs\") pod \"placement-5c4946988b-h259p\" (UID: \"342978a1-dc32-4347-bf34-6782c201f033\") " pod="openstack/placement-5c4946988b-h259p" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.932509 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/342978a1-dc32-4347-bf34-6782c201f033-scripts\") pod \"placement-5c4946988b-h259p\" (UID: \"342978a1-dc32-4347-bf34-6782c201f033\") " pod="openstack/placement-5c4946988b-h259p" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.933579 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/342978a1-dc32-4347-bf34-6782c201f033-public-tls-certs\") pod \"placement-5c4946988b-h259p\" (UID: \"342978a1-dc32-4347-bf34-6782c201f033\") " pod="openstack/placement-5c4946988b-h259p" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.934882 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/342978a1-dc32-4347-bf34-6782c201f033-internal-tls-certs\") pod \"placement-5c4946988b-h259p\" (UID: \"342978a1-dc32-4347-bf34-6782c201f033\") " pod="openstack/placement-5c4946988b-h259p" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.935464 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/342978a1-dc32-4347-bf34-6782c201f033-config-data\") pod \"placement-5c4946988b-h259p\" (UID: \"342978a1-dc32-4347-bf34-6782c201f033\") " pod="openstack/placement-5c4946988b-h259p" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.942495 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/342978a1-dc32-4347-bf34-6782c201f033-combined-ca-bundle\") pod \"placement-5c4946988b-h259p\" (UID: \"342978a1-dc32-4347-bf34-6782c201f033\") " pod="openstack/placement-5c4946988b-h259p" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.953876 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvvgp\" (UniqueName: \"kubernetes.io/projected/342978a1-dc32-4347-bf34-6782c201f033-kube-api-access-zvvgp\") pod \"placement-5c4946988b-h259p\" (UID: \"342978a1-dc32-4347-bf34-6782c201f033\") " pod="openstack/placement-5c4946988b-h259p" Oct 07 14:03:30 crc kubenswrapper[4959]: I1007 14:03:30.993402 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5c4946988b-h259p" Oct 07 14:03:31 crc kubenswrapper[4959]: I1007 14:03:31.461030 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5c4946988b-h259p"] Oct 07 14:03:31 crc kubenswrapper[4959]: W1007 14:03:31.475375 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod342978a1_dc32_4347_bf34_6782c201f033.slice/crio-20c7068f6f7a08f7bace350fe00f7c4ce8fca6293ff0ec23e146681642127eb0 WatchSource:0}: Error finding container 20c7068f6f7a08f7bace350fe00f7c4ce8fca6293ff0ec23e146681642127eb0: Status 404 returned error can't find the container with id 20c7068f6f7a08f7bace350fe00f7c4ce8fca6293ff0ec23e146681642127eb0 Oct 07 14:03:31 crc kubenswrapper[4959]: I1007 14:03:31.494790 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5c4946988b-h259p" event={"ID":"342978a1-dc32-4347-bf34-6782c201f033","Type":"ContainerStarted","Data":"20c7068f6f7a08f7bace350fe00f7c4ce8fca6293ff0ec23e146681642127eb0"} Oct 07 14:03:31 crc kubenswrapper[4959]: I1007 14:03:31.497410 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5c77f969b5-4w2xs" event={"ID":"33e694c5-266c-40e1-a805-c174cd094645","Type":"ContainerStarted","Data":"34d259feba1bf937def4c3a5a556ee6b4d8b43e291002b6f3280c8843d7ad0a2"} Oct 07 14:03:31 crc kubenswrapper[4959]: I1007 14:03:31.513088 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5c77f969b5-4w2xs" podStartSLOduration=2.513073345 podStartE2EDuration="2.513073345s" podCreationTimestamp="2025-10-07 14:03:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:03:31.512535062 +0000 UTC m=+1093.595939387" watchObservedRunningTime="2025-10-07 14:03:31.513073345 +0000 UTC m=+1093.596477670" Oct 07 14:03:31 crc kubenswrapper[4959]: I1007 14:03:31.772306 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2hhl9" Oct 07 14:03:31 crc kubenswrapper[4959]: I1007 14:03:31.948918 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hw67n\" (UniqueName: \"kubernetes.io/projected/c3b8b961-3c0e-47dc-b28f-d55c9821ab34-kube-api-access-hw67n\") pod \"c3b8b961-3c0e-47dc-b28f-d55c9821ab34\" (UID: \"c3b8b961-3c0e-47dc-b28f-d55c9821ab34\") " Oct 07 14:03:31 crc kubenswrapper[4959]: I1007 14:03:31.949458 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c3b8b961-3c0e-47dc-b28f-d55c9821ab34-db-sync-config-data\") pod \"c3b8b961-3c0e-47dc-b28f-d55c9821ab34\" (UID: \"c3b8b961-3c0e-47dc-b28f-d55c9821ab34\") " Oct 07 14:03:31 crc kubenswrapper[4959]: I1007 14:03:31.949574 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3b8b961-3c0e-47dc-b28f-d55c9821ab34-combined-ca-bundle\") pod \"c3b8b961-3c0e-47dc-b28f-d55c9821ab34\" (UID: \"c3b8b961-3c0e-47dc-b28f-d55c9821ab34\") " Oct 07 14:03:31 crc kubenswrapper[4959]: I1007 14:03:31.953592 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3b8b961-3c0e-47dc-b28f-d55c9821ab34-kube-api-access-hw67n" (OuterVolumeSpecName: "kube-api-access-hw67n") pod "c3b8b961-3c0e-47dc-b28f-d55c9821ab34" (UID: "c3b8b961-3c0e-47dc-b28f-d55c9821ab34"). InnerVolumeSpecName "kube-api-access-hw67n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:03:31 crc kubenswrapper[4959]: I1007 14:03:31.953858 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3b8b961-3c0e-47dc-b28f-d55c9821ab34-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "c3b8b961-3c0e-47dc-b28f-d55c9821ab34" (UID: "c3b8b961-3c0e-47dc-b28f-d55c9821ab34"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:31 crc kubenswrapper[4959]: I1007 14:03:31.987981 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3b8b961-3c0e-47dc-b28f-d55c9821ab34-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c3b8b961-3c0e-47dc-b28f-d55c9821ab34" (UID: "c3b8b961-3c0e-47dc-b28f-d55c9821ab34"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.051329 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3b8b961-3c0e-47dc-b28f-d55c9821ab34-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.051372 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hw67n\" (UniqueName: \"kubernetes.io/projected/c3b8b961-3c0e-47dc-b28f-d55c9821ab34-kube-api-access-hw67n\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.051384 4959 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c3b8b961-3c0e-47dc-b28f-d55c9821ab34-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.506992 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5c4946988b-h259p" event={"ID":"342978a1-dc32-4347-bf34-6782c201f033","Type":"ContainerStarted","Data":"bcd7f1828a90487b319e75b182e02508f4ac0f22bb499cce90c07380cac3da87"} Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.509432 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2hhl9" event={"ID":"c3b8b961-3c0e-47dc-b28f-d55c9821ab34","Type":"ContainerDied","Data":"9b40f4a3a03e61fc71e06571fa3dffb62281881bb7810b6d9fed9638f36c7a3d"} Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.509458 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b40f4a3a03e61fc71e06571fa3dffb62281881bb7810b6d9fed9638f36c7a3d" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.509463 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2hhl9" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.509563 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5c77f969b5-4w2xs" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.766165 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-76c697766b-9qfh8"] Oct 07 14:03:32 crc kubenswrapper[4959]: E1007 14:03:32.766537 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3b8b961-3c0e-47dc-b28f-d55c9821ab34" containerName="barbican-db-sync" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.766552 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3b8b961-3c0e-47dc-b28f-d55c9821ab34" containerName="barbican-db-sync" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.766727 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3b8b961-3c0e-47dc-b28f-d55c9821ab34" containerName="barbican-db-sync" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.767629 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-76c697766b-9qfh8" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.771666 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-tgdwf" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.773275 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.773450 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.773818 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4a56ff7-04cc-48e9-be4a-651a98c06204-config-data\") pod \"barbican-keystone-listener-76c697766b-9qfh8\" (UID: \"a4a56ff7-04cc-48e9-be4a-651a98c06204\") " pod="openstack/barbican-keystone-listener-76c697766b-9qfh8" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.773906 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4a56ff7-04cc-48e9-be4a-651a98c06204-combined-ca-bundle\") pod \"barbican-keystone-listener-76c697766b-9qfh8\" (UID: \"a4a56ff7-04cc-48e9-be4a-651a98c06204\") " pod="openstack/barbican-keystone-listener-76c697766b-9qfh8" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.773956 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pb29\" (UniqueName: \"kubernetes.io/projected/a4a56ff7-04cc-48e9-be4a-651a98c06204-kube-api-access-2pb29\") pod \"barbican-keystone-listener-76c697766b-9qfh8\" (UID: \"a4a56ff7-04cc-48e9-be4a-651a98c06204\") " pod="openstack/barbican-keystone-listener-76c697766b-9qfh8" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.774027 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4a56ff7-04cc-48e9-be4a-651a98c06204-logs\") pod \"barbican-keystone-listener-76c697766b-9qfh8\" (UID: \"a4a56ff7-04cc-48e9-be4a-651a98c06204\") " pod="openstack/barbican-keystone-listener-76c697766b-9qfh8" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.774080 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a4a56ff7-04cc-48e9-be4a-651a98c06204-config-data-custom\") pod \"barbican-keystone-listener-76c697766b-9qfh8\" (UID: \"a4a56ff7-04cc-48e9-be4a-651a98c06204\") " pod="openstack/barbican-keystone-listener-76c697766b-9qfh8" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.779296 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7748b8ffbf-c9js9"] Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.788145 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7748b8ffbf-c9js9" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.792907 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-76c697766b-9qfh8"] Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.797696 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.801776 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7748b8ffbf-c9js9"] Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.859765 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f84c94c5-cbptf"] Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.861232 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f84c94c5-cbptf" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.877839 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4a56ff7-04cc-48e9-be4a-651a98c06204-combined-ca-bundle\") pod \"barbican-keystone-listener-76c697766b-9qfh8\" (UID: \"a4a56ff7-04cc-48e9-be4a-651a98c06204\") " pod="openstack/barbican-keystone-listener-76c697766b-9qfh8" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.878182 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pb29\" (UniqueName: \"kubernetes.io/projected/a4a56ff7-04cc-48e9-be4a-651a98c06204-kube-api-access-2pb29\") pod \"barbican-keystone-listener-76c697766b-9qfh8\" (UID: \"a4a56ff7-04cc-48e9-be4a-651a98c06204\") " pod="openstack/barbican-keystone-listener-76c697766b-9qfh8" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.878257 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4a56ff7-04cc-48e9-be4a-651a98c06204-logs\") pod \"barbican-keystone-listener-76c697766b-9qfh8\" (UID: \"a4a56ff7-04cc-48e9-be4a-651a98c06204\") " pod="openstack/barbican-keystone-listener-76c697766b-9qfh8" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.878293 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a4a56ff7-04cc-48e9-be4a-651a98c06204-config-data-custom\") pod \"barbican-keystone-listener-76c697766b-9qfh8\" (UID: \"a4a56ff7-04cc-48e9-be4a-651a98c06204\") " pod="openstack/barbican-keystone-listener-76c697766b-9qfh8" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.878832 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4a56ff7-04cc-48e9-be4a-651a98c06204-config-data\") pod \"barbican-keystone-listener-76c697766b-9qfh8\" (UID: \"a4a56ff7-04cc-48e9-be4a-651a98c06204\") " pod="openstack/barbican-keystone-listener-76c697766b-9qfh8" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.889109 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f84c94c5-cbptf"] Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.903616 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a4a56ff7-04cc-48e9-be4a-651a98c06204-config-data-custom\") pod \"barbican-keystone-listener-76c697766b-9qfh8\" (UID: \"a4a56ff7-04cc-48e9-be4a-651a98c06204\") " pod="openstack/barbican-keystone-listener-76c697766b-9qfh8" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.903620 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pb29\" (UniqueName: \"kubernetes.io/projected/a4a56ff7-04cc-48e9-be4a-651a98c06204-kube-api-access-2pb29\") pod \"barbican-keystone-listener-76c697766b-9qfh8\" (UID: \"a4a56ff7-04cc-48e9-be4a-651a98c06204\") " pod="openstack/barbican-keystone-listener-76c697766b-9qfh8" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.903791 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4a56ff7-04cc-48e9-be4a-651a98c06204-logs\") pod \"barbican-keystone-listener-76c697766b-9qfh8\" (UID: \"a4a56ff7-04cc-48e9-be4a-651a98c06204\") " pod="openstack/barbican-keystone-listener-76c697766b-9qfh8" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.904334 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4a56ff7-04cc-48e9-be4a-651a98c06204-combined-ca-bundle\") pod \"barbican-keystone-listener-76c697766b-9qfh8\" (UID: \"a4a56ff7-04cc-48e9-be4a-651a98c06204\") " pod="openstack/barbican-keystone-listener-76c697766b-9qfh8" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.909734 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4a56ff7-04cc-48e9-be4a-651a98c06204-config-data\") pod \"barbican-keystone-listener-76c697766b-9qfh8\" (UID: \"a4a56ff7-04cc-48e9-be4a-651a98c06204\") " pod="openstack/barbican-keystone-listener-76c697766b-9qfh8" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.920555 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-56c8db7dd-6jq82"] Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.921929 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-56c8db7dd-6jq82" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.929434 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.936216 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-56c8db7dd-6jq82"] Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.980814 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-dns-svc\") pod \"dnsmasq-dns-6f84c94c5-cbptf\" (UID: \"541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa\") " pod="openstack/dnsmasq-dns-6f84c94c5-cbptf" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.980854 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-ovsdbserver-sb\") pod \"dnsmasq-dns-6f84c94c5-cbptf\" (UID: \"541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa\") " pod="openstack/dnsmasq-dns-6f84c94c5-cbptf" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.980890 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-ovsdbserver-nb\") pod \"dnsmasq-dns-6f84c94c5-cbptf\" (UID: \"541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa\") " pod="openstack/dnsmasq-dns-6f84c94c5-cbptf" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.980938 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zgcj\" (UniqueName: \"kubernetes.io/projected/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-kube-api-access-9zgcj\") pod \"dnsmasq-dns-6f84c94c5-cbptf\" (UID: \"541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa\") " pod="openstack/dnsmasq-dns-6f84c94c5-cbptf" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.980976 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4f04734-09dc-47cf-9395-5f2ca7739a3f-combined-ca-bundle\") pod \"barbican-worker-7748b8ffbf-c9js9\" (UID: \"d4f04734-09dc-47cf-9395-5f2ca7739a3f\") " pod="openstack/barbican-worker-7748b8ffbf-c9js9" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.981014 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4f04734-09dc-47cf-9395-5f2ca7739a3f-logs\") pod \"barbican-worker-7748b8ffbf-c9js9\" (UID: \"d4f04734-09dc-47cf-9395-5f2ca7739a3f\") " pod="openstack/barbican-worker-7748b8ffbf-c9js9" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.981068 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4f04734-09dc-47cf-9395-5f2ca7739a3f-config-data-custom\") pod \"barbican-worker-7748b8ffbf-c9js9\" (UID: \"d4f04734-09dc-47cf-9395-5f2ca7739a3f\") " pod="openstack/barbican-worker-7748b8ffbf-c9js9" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.981123 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-config\") pod \"dnsmasq-dns-6f84c94c5-cbptf\" (UID: \"541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa\") " pod="openstack/dnsmasq-dns-6f84c94c5-cbptf" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.981155 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdb55\" (UniqueName: \"kubernetes.io/projected/d4f04734-09dc-47cf-9395-5f2ca7739a3f-kube-api-access-qdb55\") pod \"barbican-worker-7748b8ffbf-c9js9\" (UID: \"d4f04734-09dc-47cf-9395-5f2ca7739a3f\") " pod="openstack/barbican-worker-7748b8ffbf-c9js9" Oct 07 14:03:32 crc kubenswrapper[4959]: I1007 14:03:32.981195 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4f04734-09dc-47cf-9395-5f2ca7739a3f-config-data\") pod \"barbican-worker-7748b8ffbf-c9js9\" (UID: \"d4f04734-09dc-47cf-9395-5f2ca7739a3f\") " pod="openstack/barbican-worker-7748b8ffbf-c9js9" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.082497 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-dns-svc\") pod \"dnsmasq-dns-6f84c94c5-cbptf\" (UID: \"541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa\") " pod="openstack/dnsmasq-dns-6f84c94c5-cbptf" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.082545 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-ovsdbserver-sb\") pod \"dnsmasq-dns-6f84c94c5-cbptf\" (UID: \"541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa\") " pod="openstack/dnsmasq-dns-6f84c94c5-cbptf" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.082570 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf98f\" (UniqueName: \"kubernetes.io/projected/36d7df61-5454-44d6-a7e1-04a0fc4cb621-kube-api-access-lf98f\") pod \"barbican-api-56c8db7dd-6jq82\" (UID: \"36d7df61-5454-44d6-a7e1-04a0fc4cb621\") " pod="openstack/barbican-api-56c8db7dd-6jq82" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.082596 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-ovsdbserver-nb\") pod \"dnsmasq-dns-6f84c94c5-cbptf\" (UID: \"541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa\") " pod="openstack/dnsmasq-dns-6f84c94c5-cbptf" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.082628 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zgcj\" (UniqueName: \"kubernetes.io/projected/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-kube-api-access-9zgcj\") pod \"dnsmasq-dns-6f84c94c5-cbptf\" (UID: \"541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa\") " pod="openstack/dnsmasq-dns-6f84c94c5-cbptf" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.082647 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36d7df61-5454-44d6-a7e1-04a0fc4cb621-logs\") pod \"barbican-api-56c8db7dd-6jq82\" (UID: \"36d7df61-5454-44d6-a7e1-04a0fc4cb621\") " pod="openstack/barbican-api-56c8db7dd-6jq82" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.082721 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4f04734-09dc-47cf-9395-5f2ca7739a3f-combined-ca-bundle\") pod \"barbican-worker-7748b8ffbf-c9js9\" (UID: \"d4f04734-09dc-47cf-9395-5f2ca7739a3f\") " pod="openstack/barbican-worker-7748b8ffbf-c9js9" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.082751 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4f04734-09dc-47cf-9395-5f2ca7739a3f-logs\") pod \"barbican-worker-7748b8ffbf-c9js9\" (UID: \"d4f04734-09dc-47cf-9395-5f2ca7739a3f\") " pod="openstack/barbican-worker-7748b8ffbf-c9js9" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.082774 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/36d7df61-5454-44d6-a7e1-04a0fc4cb621-config-data-custom\") pod \"barbican-api-56c8db7dd-6jq82\" (UID: \"36d7df61-5454-44d6-a7e1-04a0fc4cb621\") " pod="openstack/barbican-api-56c8db7dd-6jq82" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.082808 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4f04734-09dc-47cf-9395-5f2ca7739a3f-config-data-custom\") pod \"barbican-worker-7748b8ffbf-c9js9\" (UID: \"d4f04734-09dc-47cf-9395-5f2ca7739a3f\") " pod="openstack/barbican-worker-7748b8ffbf-c9js9" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.082834 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-config\") pod \"dnsmasq-dns-6f84c94c5-cbptf\" (UID: \"541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa\") " pod="openstack/dnsmasq-dns-6f84c94c5-cbptf" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.082850 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36d7df61-5454-44d6-a7e1-04a0fc4cb621-config-data\") pod \"barbican-api-56c8db7dd-6jq82\" (UID: \"36d7df61-5454-44d6-a7e1-04a0fc4cb621\") " pod="openstack/barbican-api-56c8db7dd-6jq82" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.082875 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdb55\" (UniqueName: \"kubernetes.io/projected/d4f04734-09dc-47cf-9395-5f2ca7739a3f-kube-api-access-qdb55\") pod \"barbican-worker-7748b8ffbf-c9js9\" (UID: \"d4f04734-09dc-47cf-9395-5f2ca7739a3f\") " pod="openstack/barbican-worker-7748b8ffbf-c9js9" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.082899 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36d7df61-5454-44d6-a7e1-04a0fc4cb621-combined-ca-bundle\") pod \"barbican-api-56c8db7dd-6jq82\" (UID: \"36d7df61-5454-44d6-a7e1-04a0fc4cb621\") " pod="openstack/barbican-api-56c8db7dd-6jq82" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.082921 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4f04734-09dc-47cf-9395-5f2ca7739a3f-config-data\") pod \"barbican-worker-7748b8ffbf-c9js9\" (UID: \"d4f04734-09dc-47cf-9395-5f2ca7739a3f\") " pod="openstack/barbican-worker-7748b8ffbf-c9js9" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.083647 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-dns-svc\") pod \"dnsmasq-dns-6f84c94c5-cbptf\" (UID: \"541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa\") " pod="openstack/dnsmasq-dns-6f84c94c5-cbptf" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.083655 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-ovsdbserver-sb\") pod \"dnsmasq-dns-6f84c94c5-cbptf\" (UID: \"541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa\") " pod="openstack/dnsmasq-dns-6f84c94c5-cbptf" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.083912 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-ovsdbserver-nb\") pod \"dnsmasq-dns-6f84c94c5-cbptf\" (UID: \"541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa\") " pod="openstack/dnsmasq-dns-6f84c94c5-cbptf" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.084118 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4f04734-09dc-47cf-9395-5f2ca7739a3f-logs\") pod \"barbican-worker-7748b8ffbf-c9js9\" (UID: \"d4f04734-09dc-47cf-9395-5f2ca7739a3f\") " pod="openstack/barbican-worker-7748b8ffbf-c9js9" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.084406 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-config\") pod \"dnsmasq-dns-6f84c94c5-cbptf\" (UID: \"541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa\") " pod="openstack/dnsmasq-dns-6f84c94c5-cbptf" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.089082 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4f04734-09dc-47cf-9395-5f2ca7739a3f-config-data-custom\") pod \"barbican-worker-7748b8ffbf-c9js9\" (UID: \"d4f04734-09dc-47cf-9395-5f2ca7739a3f\") " pod="openstack/barbican-worker-7748b8ffbf-c9js9" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.102199 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4f04734-09dc-47cf-9395-5f2ca7739a3f-combined-ca-bundle\") pod \"barbican-worker-7748b8ffbf-c9js9\" (UID: \"d4f04734-09dc-47cf-9395-5f2ca7739a3f\") " pod="openstack/barbican-worker-7748b8ffbf-c9js9" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.104603 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-76c697766b-9qfh8" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.105234 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4f04734-09dc-47cf-9395-5f2ca7739a3f-config-data\") pod \"barbican-worker-7748b8ffbf-c9js9\" (UID: \"d4f04734-09dc-47cf-9395-5f2ca7739a3f\") " pod="openstack/barbican-worker-7748b8ffbf-c9js9" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.105433 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdb55\" (UniqueName: \"kubernetes.io/projected/d4f04734-09dc-47cf-9395-5f2ca7739a3f-kube-api-access-qdb55\") pod \"barbican-worker-7748b8ffbf-c9js9\" (UID: \"d4f04734-09dc-47cf-9395-5f2ca7739a3f\") " pod="openstack/barbican-worker-7748b8ffbf-c9js9" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.105740 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zgcj\" (UniqueName: \"kubernetes.io/projected/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-kube-api-access-9zgcj\") pod \"dnsmasq-dns-6f84c94c5-cbptf\" (UID: \"541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa\") " pod="openstack/dnsmasq-dns-6f84c94c5-cbptf" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.129384 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7748b8ffbf-c9js9" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.184779 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/36d7df61-5454-44d6-a7e1-04a0fc4cb621-config-data-custom\") pod \"barbican-api-56c8db7dd-6jq82\" (UID: \"36d7df61-5454-44d6-a7e1-04a0fc4cb621\") " pod="openstack/barbican-api-56c8db7dd-6jq82" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.184862 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36d7df61-5454-44d6-a7e1-04a0fc4cb621-config-data\") pod \"barbican-api-56c8db7dd-6jq82\" (UID: \"36d7df61-5454-44d6-a7e1-04a0fc4cb621\") " pod="openstack/barbican-api-56c8db7dd-6jq82" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.184905 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36d7df61-5454-44d6-a7e1-04a0fc4cb621-combined-ca-bundle\") pod \"barbican-api-56c8db7dd-6jq82\" (UID: \"36d7df61-5454-44d6-a7e1-04a0fc4cb621\") " pod="openstack/barbican-api-56c8db7dd-6jq82" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.184941 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf98f\" (UniqueName: \"kubernetes.io/projected/36d7df61-5454-44d6-a7e1-04a0fc4cb621-kube-api-access-lf98f\") pod \"barbican-api-56c8db7dd-6jq82\" (UID: \"36d7df61-5454-44d6-a7e1-04a0fc4cb621\") " pod="openstack/barbican-api-56c8db7dd-6jq82" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.184983 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36d7df61-5454-44d6-a7e1-04a0fc4cb621-logs\") pod \"barbican-api-56c8db7dd-6jq82\" (UID: \"36d7df61-5454-44d6-a7e1-04a0fc4cb621\") " pod="openstack/barbican-api-56c8db7dd-6jq82" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.185420 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36d7df61-5454-44d6-a7e1-04a0fc4cb621-logs\") pod \"barbican-api-56c8db7dd-6jq82\" (UID: \"36d7df61-5454-44d6-a7e1-04a0fc4cb621\") " pod="openstack/barbican-api-56c8db7dd-6jq82" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.188593 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36d7df61-5454-44d6-a7e1-04a0fc4cb621-config-data\") pod \"barbican-api-56c8db7dd-6jq82\" (UID: \"36d7df61-5454-44d6-a7e1-04a0fc4cb621\") " pod="openstack/barbican-api-56c8db7dd-6jq82" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.189758 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36d7df61-5454-44d6-a7e1-04a0fc4cb621-combined-ca-bundle\") pod \"barbican-api-56c8db7dd-6jq82\" (UID: \"36d7df61-5454-44d6-a7e1-04a0fc4cb621\") " pod="openstack/barbican-api-56c8db7dd-6jq82" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.191603 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/36d7df61-5454-44d6-a7e1-04a0fc4cb621-config-data-custom\") pod \"barbican-api-56c8db7dd-6jq82\" (UID: \"36d7df61-5454-44d6-a7e1-04a0fc4cb621\") " pod="openstack/barbican-api-56c8db7dd-6jq82" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.198787 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f84c94c5-cbptf" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.200812 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf98f\" (UniqueName: \"kubernetes.io/projected/36d7df61-5454-44d6-a7e1-04a0fc4cb621-kube-api-access-lf98f\") pod \"barbican-api-56c8db7dd-6jq82\" (UID: \"36d7df61-5454-44d6-a7e1-04a0fc4cb621\") " pod="openstack/barbican-api-56c8db7dd-6jq82" Oct 07 14:03:33 crc kubenswrapper[4959]: I1007 14:03:33.286257 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-56c8db7dd-6jq82" Oct 07 14:03:34 crc kubenswrapper[4959]: I1007 14:03:34.534698 4959 generic.go:334] "Generic (PLEG): container finished" podID="806fce72-373c-4c34-8293-3051691eb55f" containerID="d0594f269876109a127823477b2fc4b883d1d61793068015b09be8b59b07c22a" exitCode=0 Oct 07 14:03:34 crc kubenswrapper[4959]: I1007 14:03:34.535008 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4sfdr" event={"ID":"806fce72-373c-4c34-8293-3051691eb55f","Type":"ContainerDied","Data":"d0594f269876109a127823477b2fc4b883d1d61793068015b09be8b59b07c22a"} Oct 07 14:03:35 crc kubenswrapper[4959]: I1007 14:03:35.370596 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-646c6bbb48-kcmnj"] Oct 07 14:03:35 crc kubenswrapper[4959]: I1007 14:03:35.371864 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-646c6bbb48-kcmnj" Oct 07 14:03:35 crc kubenswrapper[4959]: I1007 14:03:35.377458 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 07 14:03:35 crc kubenswrapper[4959]: I1007 14:03:35.377951 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 07 14:03:35 crc kubenswrapper[4959]: I1007 14:03:35.384924 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-646c6bbb48-kcmnj"] Oct 07 14:03:35 crc kubenswrapper[4959]: I1007 14:03:35.526426 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a0d41b9-b996-4597-8646-4fb61c6dc354-logs\") pod \"barbican-api-646c6bbb48-kcmnj\" (UID: \"3a0d41b9-b996-4597-8646-4fb61c6dc354\") " pod="openstack/barbican-api-646c6bbb48-kcmnj" Oct 07 14:03:35 crc kubenswrapper[4959]: I1007 14:03:35.526536 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a0d41b9-b996-4597-8646-4fb61c6dc354-public-tls-certs\") pod \"barbican-api-646c6bbb48-kcmnj\" (UID: \"3a0d41b9-b996-4597-8646-4fb61c6dc354\") " pod="openstack/barbican-api-646c6bbb48-kcmnj" Oct 07 14:03:35 crc kubenswrapper[4959]: I1007 14:03:35.526625 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a0d41b9-b996-4597-8646-4fb61c6dc354-config-data\") pod \"barbican-api-646c6bbb48-kcmnj\" (UID: \"3a0d41b9-b996-4597-8646-4fb61c6dc354\") " pod="openstack/barbican-api-646c6bbb48-kcmnj" Oct 07 14:03:35 crc kubenswrapper[4959]: I1007 14:03:35.526737 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-877lj\" (UniqueName: \"kubernetes.io/projected/3a0d41b9-b996-4597-8646-4fb61c6dc354-kube-api-access-877lj\") pod \"barbican-api-646c6bbb48-kcmnj\" (UID: \"3a0d41b9-b996-4597-8646-4fb61c6dc354\") " pod="openstack/barbican-api-646c6bbb48-kcmnj" Oct 07 14:03:35 crc kubenswrapper[4959]: I1007 14:03:35.526803 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a0d41b9-b996-4597-8646-4fb61c6dc354-config-data-custom\") pod \"barbican-api-646c6bbb48-kcmnj\" (UID: \"3a0d41b9-b996-4597-8646-4fb61c6dc354\") " pod="openstack/barbican-api-646c6bbb48-kcmnj" Oct 07 14:03:35 crc kubenswrapper[4959]: I1007 14:03:35.526831 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a0d41b9-b996-4597-8646-4fb61c6dc354-combined-ca-bundle\") pod \"barbican-api-646c6bbb48-kcmnj\" (UID: \"3a0d41b9-b996-4597-8646-4fb61c6dc354\") " pod="openstack/barbican-api-646c6bbb48-kcmnj" Oct 07 14:03:35 crc kubenswrapper[4959]: I1007 14:03:35.526935 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a0d41b9-b996-4597-8646-4fb61c6dc354-internal-tls-certs\") pod \"barbican-api-646c6bbb48-kcmnj\" (UID: \"3a0d41b9-b996-4597-8646-4fb61c6dc354\") " pod="openstack/barbican-api-646c6bbb48-kcmnj" Oct 07 14:03:35 crc kubenswrapper[4959]: I1007 14:03:35.543875 4959 generic.go:334] "Generic (PLEG): container finished" podID="7b70a321-ffb7-429f-b825-4cd872ded9ff" containerID="1c98cd878dfdc18d00933f385480ff055b149f850fe48ee3960f16c3ce508cb7" exitCode=0 Oct 07 14:03:35 crc kubenswrapper[4959]: I1007 14:03:35.543967 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-jfrdg" event={"ID":"7b70a321-ffb7-429f-b825-4cd872ded9ff","Type":"ContainerDied","Data":"1c98cd878dfdc18d00933f385480ff055b149f850fe48ee3960f16c3ce508cb7"} Oct 07 14:03:35 crc kubenswrapper[4959]: I1007 14:03:35.633448 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a0d41b9-b996-4597-8646-4fb61c6dc354-internal-tls-certs\") pod \"barbican-api-646c6bbb48-kcmnj\" (UID: \"3a0d41b9-b996-4597-8646-4fb61c6dc354\") " pod="openstack/barbican-api-646c6bbb48-kcmnj" Oct 07 14:03:35 crc kubenswrapper[4959]: I1007 14:03:35.633522 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a0d41b9-b996-4597-8646-4fb61c6dc354-logs\") pod \"barbican-api-646c6bbb48-kcmnj\" (UID: \"3a0d41b9-b996-4597-8646-4fb61c6dc354\") " pod="openstack/barbican-api-646c6bbb48-kcmnj" Oct 07 14:03:35 crc kubenswrapper[4959]: I1007 14:03:35.633551 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a0d41b9-b996-4597-8646-4fb61c6dc354-public-tls-certs\") pod \"barbican-api-646c6bbb48-kcmnj\" (UID: \"3a0d41b9-b996-4597-8646-4fb61c6dc354\") " pod="openstack/barbican-api-646c6bbb48-kcmnj" Oct 07 14:03:35 crc kubenswrapper[4959]: I1007 14:03:35.633578 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a0d41b9-b996-4597-8646-4fb61c6dc354-config-data\") pod \"barbican-api-646c6bbb48-kcmnj\" (UID: \"3a0d41b9-b996-4597-8646-4fb61c6dc354\") " pod="openstack/barbican-api-646c6bbb48-kcmnj" Oct 07 14:03:35 crc kubenswrapper[4959]: I1007 14:03:35.633617 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-877lj\" (UniqueName: \"kubernetes.io/projected/3a0d41b9-b996-4597-8646-4fb61c6dc354-kube-api-access-877lj\") pod \"barbican-api-646c6bbb48-kcmnj\" (UID: \"3a0d41b9-b996-4597-8646-4fb61c6dc354\") " pod="openstack/barbican-api-646c6bbb48-kcmnj" Oct 07 14:03:35 crc kubenswrapper[4959]: I1007 14:03:35.633642 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a0d41b9-b996-4597-8646-4fb61c6dc354-config-data-custom\") pod \"barbican-api-646c6bbb48-kcmnj\" (UID: \"3a0d41b9-b996-4597-8646-4fb61c6dc354\") " pod="openstack/barbican-api-646c6bbb48-kcmnj" Oct 07 14:03:35 crc kubenswrapper[4959]: I1007 14:03:35.633675 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a0d41b9-b996-4597-8646-4fb61c6dc354-combined-ca-bundle\") pod \"barbican-api-646c6bbb48-kcmnj\" (UID: \"3a0d41b9-b996-4597-8646-4fb61c6dc354\") " pod="openstack/barbican-api-646c6bbb48-kcmnj" Oct 07 14:03:35 crc kubenswrapper[4959]: I1007 14:03:35.637053 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a0d41b9-b996-4597-8646-4fb61c6dc354-combined-ca-bundle\") pod \"barbican-api-646c6bbb48-kcmnj\" (UID: \"3a0d41b9-b996-4597-8646-4fb61c6dc354\") " pod="openstack/barbican-api-646c6bbb48-kcmnj" Oct 07 14:03:35 crc kubenswrapper[4959]: I1007 14:03:35.638163 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a0d41b9-b996-4597-8646-4fb61c6dc354-config-data\") pod \"barbican-api-646c6bbb48-kcmnj\" (UID: \"3a0d41b9-b996-4597-8646-4fb61c6dc354\") " pod="openstack/barbican-api-646c6bbb48-kcmnj" Oct 07 14:03:35 crc kubenswrapper[4959]: I1007 14:03:35.638425 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a0d41b9-b996-4597-8646-4fb61c6dc354-logs\") pod \"barbican-api-646c6bbb48-kcmnj\" (UID: \"3a0d41b9-b996-4597-8646-4fb61c6dc354\") " pod="openstack/barbican-api-646c6bbb48-kcmnj" Oct 07 14:03:35 crc kubenswrapper[4959]: I1007 14:03:35.639771 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a0d41b9-b996-4597-8646-4fb61c6dc354-public-tls-certs\") pod \"barbican-api-646c6bbb48-kcmnj\" (UID: \"3a0d41b9-b996-4597-8646-4fb61c6dc354\") " pod="openstack/barbican-api-646c6bbb48-kcmnj" Oct 07 14:03:35 crc kubenswrapper[4959]: I1007 14:03:35.640607 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a0d41b9-b996-4597-8646-4fb61c6dc354-config-data-custom\") pod \"barbican-api-646c6bbb48-kcmnj\" (UID: \"3a0d41b9-b996-4597-8646-4fb61c6dc354\") " pod="openstack/barbican-api-646c6bbb48-kcmnj" Oct 07 14:03:35 crc kubenswrapper[4959]: I1007 14:03:35.652446 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a0d41b9-b996-4597-8646-4fb61c6dc354-internal-tls-certs\") pod \"barbican-api-646c6bbb48-kcmnj\" (UID: \"3a0d41b9-b996-4597-8646-4fb61c6dc354\") " pod="openstack/barbican-api-646c6bbb48-kcmnj" Oct 07 14:03:35 crc kubenswrapper[4959]: I1007 14:03:35.656263 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-877lj\" (UniqueName: \"kubernetes.io/projected/3a0d41b9-b996-4597-8646-4fb61c6dc354-kube-api-access-877lj\") pod \"barbican-api-646c6bbb48-kcmnj\" (UID: \"3a0d41b9-b996-4597-8646-4fb61c6dc354\") " pod="openstack/barbican-api-646c6bbb48-kcmnj" Oct 07 14:03:35 crc kubenswrapper[4959]: I1007 14:03:35.708920 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-646c6bbb48-kcmnj" Oct 07 14:03:36 crc kubenswrapper[4959]: I1007 14:03:36.487965 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:03:36 crc kubenswrapper[4959]: I1007 14:03:36.488257 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:03:36 crc kubenswrapper[4959]: I1007 14:03:36.609666 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-76d6fc6cbd-2g7sq" Oct 07 14:03:36 crc kubenswrapper[4959]: I1007 14:03:36.609786 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-76d6fc6cbd-2g7sq" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.261731 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4sfdr" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.296416 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-jfrdg" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.369973 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b70a321-ffb7-429f-b825-4cd872ded9ff-config-data\") pod \"7b70a321-ffb7-429f-b825-4cd872ded9ff\" (UID: \"7b70a321-ffb7-429f-b825-4cd872ded9ff\") " Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.370033 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmzf9\" (UniqueName: \"kubernetes.io/projected/7b70a321-ffb7-429f-b825-4cd872ded9ff-kube-api-access-zmzf9\") pod \"7b70a321-ffb7-429f-b825-4cd872ded9ff\" (UID: \"7b70a321-ffb7-429f-b825-4cd872ded9ff\") " Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.370061 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/806fce72-373c-4c34-8293-3051691eb55f-combined-ca-bundle\") pod \"806fce72-373c-4c34-8293-3051691eb55f\" (UID: \"806fce72-373c-4c34-8293-3051691eb55f\") " Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.370093 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/806fce72-373c-4c34-8293-3051691eb55f-config\") pod \"806fce72-373c-4c34-8293-3051691eb55f\" (UID: \"806fce72-373c-4c34-8293-3051691eb55f\") " Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.370167 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7b70a321-ffb7-429f-b825-4cd872ded9ff-db-sync-config-data\") pod \"7b70a321-ffb7-429f-b825-4cd872ded9ff\" (UID: \"7b70a321-ffb7-429f-b825-4cd872ded9ff\") " Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.370239 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b70a321-ffb7-429f-b825-4cd872ded9ff-combined-ca-bundle\") pod \"7b70a321-ffb7-429f-b825-4cd872ded9ff\" (UID: \"7b70a321-ffb7-429f-b825-4cd872ded9ff\") " Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.370302 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b70a321-ffb7-429f-b825-4cd872ded9ff-scripts\") pod \"7b70a321-ffb7-429f-b825-4cd872ded9ff\" (UID: \"7b70a321-ffb7-429f-b825-4cd872ded9ff\") " Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.370334 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7b70a321-ffb7-429f-b825-4cd872ded9ff-etc-machine-id\") pod \"7b70a321-ffb7-429f-b825-4cd872ded9ff\" (UID: \"7b70a321-ffb7-429f-b825-4cd872ded9ff\") " Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.370365 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjmfq\" (UniqueName: \"kubernetes.io/projected/806fce72-373c-4c34-8293-3051691eb55f-kube-api-access-hjmfq\") pod \"806fce72-373c-4c34-8293-3051691eb55f\" (UID: \"806fce72-373c-4c34-8293-3051691eb55f\") " Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.382240 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b70a321-ffb7-429f-b825-4cd872ded9ff-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "7b70a321-ffb7-429f-b825-4cd872ded9ff" (UID: "7b70a321-ffb7-429f-b825-4cd872ded9ff"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.392734 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b70a321-ffb7-429f-b825-4cd872ded9ff-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "7b70a321-ffb7-429f-b825-4cd872ded9ff" (UID: "7b70a321-ffb7-429f-b825-4cd872ded9ff"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.395882 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b70a321-ffb7-429f-b825-4cd872ded9ff-scripts" (OuterVolumeSpecName: "scripts") pod "7b70a321-ffb7-429f-b825-4cd872ded9ff" (UID: "7b70a321-ffb7-429f-b825-4cd872ded9ff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:37 crc kubenswrapper[4959]: E1007 14:03:37.403585 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="cd33b52c-3f7f-425e-aa42-7342343067a6" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.418209 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/806fce72-373c-4c34-8293-3051691eb55f-kube-api-access-hjmfq" (OuterVolumeSpecName: "kube-api-access-hjmfq") pod "806fce72-373c-4c34-8293-3051691eb55f" (UID: "806fce72-373c-4c34-8293-3051691eb55f"). InnerVolumeSpecName "kube-api-access-hjmfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.418680 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b70a321-ffb7-429f-b825-4cd872ded9ff-kube-api-access-zmzf9" (OuterVolumeSpecName: "kube-api-access-zmzf9") pod "7b70a321-ffb7-429f-b825-4cd872ded9ff" (UID: "7b70a321-ffb7-429f-b825-4cd872ded9ff"). InnerVolumeSpecName "kube-api-access-zmzf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.438764 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/806fce72-373c-4c34-8293-3051691eb55f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "806fce72-373c-4c34-8293-3051691eb55f" (UID: "806fce72-373c-4c34-8293-3051691eb55f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.441651 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/806fce72-373c-4c34-8293-3051691eb55f-config" (OuterVolumeSpecName: "config") pod "806fce72-373c-4c34-8293-3051691eb55f" (UID: "806fce72-373c-4c34-8293-3051691eb55f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.458963 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-646c6bbb48-kcmnj"] Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.473054 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmzf9\" (UniqueName: \"kubernetes.io/projected/7b70a321-ffb7-429f-b825-4cd872ded9ff-kube-api-access-zmzf9\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.473080 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/806fce72-373c-4c34-8293-3051691eb55f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.473091 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/806fce72-373c-4c34-8293-3051691eb55f-config\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.473118 4959 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7b70a321-ffb7-429f-b825-4cd872ded9ff-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.473127 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b70a321-ffb7-429f-b825-4cd872ded9ff-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.473134 4959 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7b70a321-ffb7-429f-b825-4cd872ded9ff-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.473142 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjmfq\" (UniqueName: \"kubernetes.io/projected/806fce72-373c-4c34-8293-3051691eb55f-kube-api-access-hjmfq\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.473769 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b70a321-ffb7-429f-b825-4cd872ded9ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7b70a321-ffb7-429f-b825-4cd872ded9ff" (UID: "7b70a321-ffb7-429f-b825-4cd872ded9ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.482326 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b70a321-ffb7-429f-b825-4cd872ded9ff-config-data" (OuterVolumeSpecName: "config-data") pod "7b70a321-ffb7-429f-b825-4cd872ded9ff" (UID: "7b70a321-ffb7-429f-b825-4cd872ded9ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.561789 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-646c6bbb48-kcmnj" event={"ID":"3a0d41b9-b996-4597-8646-4fb61c6dc354","Type":"ContainerStarted","Data":"aaa5820997d0e11bc69a431448b7859049f82518d4e8780347659ffb48521583"} Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.563466 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-jfrdg" event={"ID":"7b70a321-ffb7-429f-b825-4cd872ded9ff","Type":"ContainerDied","Data":"5770a8e64286b3347c919d50f8ccdd718c75c59f7e590a2b8c08aac30333c6ea"} Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.563494 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-jfrdg" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.563504 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5770a8e64286b3347c919d50f8ccdd718c75c59f7e590a2b8c08aac30333c6ea" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.565813 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4sfdr" event={"ID":"806fce72-373c-4c34-8293-3051691eb55f","Type":"ContainerDied","Data":"51270a802483bad43e9e5b3f35f56ee3aaaa4587185b43816ee5320d94406ec0"} Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.565856 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51270a802483bad43e9e5b3f35f56ee3aaaa4587185b43816ee5320d94406ec0" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.565925 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4sfdr" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.569596 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5c4946988b-h259p" event={"ID":"342978a1-dc32-4347-bf34-6782c201f033","Type":"ContainerStarted","Data":"a9ca014157ad99e68fffbb10335cca077a15d4528de85686c8bac58000ffba1b"} Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.569922 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5c4946988b-h259p" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.569953 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5c4946988b-h259p" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.572811 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/placement-5c4946988b-h259p" podUID="342978a1-dc32-4347-bf34-6782c201f033" containerName="placement-log" probeResult="failure" output="Get \"https://10.217.0.147:8778/\": dial tcp 10.217.0.147:8778: connect: connection refused" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.574070 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cd33b52c-3f7f-425e-aa42-7342343067a6" containerName="ceilometer-notification-agent" containerID="cri-o://a7768743c10e8cceb2e9f59d7ad327c68847c768df2193deb859955e0f76a5de" gracePeriod=30 Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.574164 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd33b52c-3f7f-425e-aa42-7342343067a6","Type":"ContainerStarted","Data":"1fce89d9c71d41fab512a1069c3e290692b1fea1e62f647256e20a6167fe5c1a"} Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.574205 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.574257 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cd33b52c-3f7f-425e-aa42-7342343067a6" containerName="proxy-httpd" containerID="cri-o://1fce89d9c71d41fab512a1069c3e290692b1fea1e62f647256e20a6167fe5c1a" gracePeriod=30 Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.574324 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cd33b52c-3f7f-425e-aa42-7342343067a6" containerName="sg-core" containerID="cri-o://1395f86ab5b523c157ccdd52e29a4d040ff47bbc2a8f66c3696b941b2aa1e1eb" gracePeriod=30 Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.576053 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b70a321-ffb7-429f-b825-4cd872ded9ff-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.576156 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b70a321-ffb7-429f-b825-4cd872ded9ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.594286 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-5c4946988b-h259p" podStartSLOduration=7.594264944 podStartE2EDuration="7.594264944s" podCreationTimestamp="2025-10-07 14:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:03:37.589981898 +0000 UTC m=+1099.673386223" watchObservedRunningTime="2025-10-07 14:03:37.594264944 +0000 UTC m=+1099.677669269" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.714603 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-56c8db7dd-6jq82"] Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.769089 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f84c94c5-cbptf"] Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.843005 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-76c697766b-9qfh8"] Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.857421 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7748b8ffbf-c9js9"] Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.956857 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 14:03:37 crc kubenswrapper[4959]: E1007 14:03:37.957544 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b70a321-ffb7-429f-b825-4cd872ded9ff" containerName="cinder-db-sync" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.957559 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b70a321-ffb7-429f-b825-4cd872ded9ff" containerName="cinder-db-sync" Oct 07 14:03:37 crc kubenswrapper[4959]: E1007 14:03:37.957601 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="806fce72-373c-4c34-8293-3051691eb55f" containerName="neutron-db-sync" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.957608 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="806fce72-373c-4c34-8293-3051691eb55f" containerName="neutron-db-sync" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.957933 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="806fce72-373c-4c34-8293-3051691eb55f" containerName="neutron-db-sync" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.957964 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b70a321-ffb7-429f-b825-4cd872ded9ff" containerName="cinder-db-sync" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.959634 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.975476 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.975903 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.976069 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-k5lq9" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.976231 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 07 14:03:37 crc kubenswrapper[4959]: I1007 14:03:37.977874 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.056525 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f84c94c5-cbptf"] Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.106950 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63c808d6-07c9-4f88-a262-447d3f291be8-config-data\") pod \"cinder-scheduler-0\" (UID: \"63c808d6-07c9-4f88-a262-447d3f291be8\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.107001 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63c808d6-07c9-4f88-a262-447d3f291be8-scripts\") pod \"cinder-scheduler-0\" (UID: \"63c808d6-07c9-4f88-a262-447d3f291be8\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.107038 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7c55\" (UniqueName: \"kubernetes.io/projected/63c808d6-07c9-4f88-a262-447d3f291be8-kube-api-access-x7c55\") pod \"cinder-scheduler-0\" (UID: \"63c808d6-07c9-4f88-a262-447d3f291be8\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.107215 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63c808d6-07c9-4f88-a262-447d3f291be8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"63c808d6-07c9-4f88-a262-447d3f291be8\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.107359 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/63c808d6-07c9-4f88-a262-447d3f291be8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"63c808d6-07c9-4f88-a262-447d3f291be8\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.107473 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/63c808d6-07c9-4f88-a262-447d3f291be8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"63c808d6-07c9-4f88-a262-447d3f291be8\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.133005 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-66996fc899-rbgs7"] Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.134432 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66996fc899-rbgs7" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.155545 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66996fc899-rbgs7"] Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.162464 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.169159 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.171553 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.182654 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.209170 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/63c808d6-07c9-4f88-a262-447d3f291be8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"63c808d6-07c9-4f88-a262-447d3f291be8\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.209221 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b48c437f-bfc5-49c8-bac3-ef4a627fc581-config\") pod \"dnsmasq-dns-66996fc899-rbgs7\" (UID: \"b48c437f-bfc5-49c8-bac3-ef4a627fc581\") " pod="openstack/dnsmasq-dns-66996fc899-rbgs7" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.209240 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b48c437f-bfc5-49c8-bac3-ef4a627fc581-ovsdbserver-sb\") pod \"dnsmasq-dns-66996fc899-rbgs7\" (UID: \"b48c437f-bfc5-49c8-bac3-ef4a627fc581\") " pod="openstack/dnsmasq-dns-66996fc899-rbgs7" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.209265 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-config-data-custom\") pod \"cinder-api-0\" (UID: \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\") " pod="openstack/cinder-api-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.209297 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\") " pod="openstack/cinder-api-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.209316 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b48c437f-bfc5-49c8-bac3-ef4a627fc581-ovsdbserver-nb\") pod \"dnsmasq-dns-66996fc899-rbgs7\" (UID: \"b48c437f-bfc5-49c8-bac3-ef4a627fc581\") " pod="openstack/dnsmasq-dns-66996fc899-rbgs7" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.209350 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\") " pod="openstack/cinder-api-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.209368 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63c808d6-07c9-4f88-a262-447d3f291be8-config-data\") pod \"cinder-scheduler-0\" (UID: \"63c808d6-07c9-4f88-a262-447d3f291be8\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.209391 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wndnc\" (UniqueName: \"kubernetes.io/projected/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-kube-api-access-wndnc\") pod \"cinder-api-0\" (UID: \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\") " pod="openstack/cinder-api-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.209407 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63c808d6-07c9-4f88-a262-447d3f291be8-scripts\") pod \"cinder-scheduler-0\" (UID: \"63c808d6-07c9-4f88-a262-447d3f291be8\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.209430 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b48c437f-bfc5-49c8-bac3-ef4a627fc581-dns-svc\") pod \"dnsmasq-dns-66996fc899-rbgs7\" (UID: \"b48c437f-bfc5-49c8-bac3-ef4a627fc581\") " pod="openstack/dnsmasq-dns-66996fc899-rbgs7" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.209451 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7c55\" (UniqueName: \"kubernetes.io/projected/63c808d6-07c9-4f88-a262-447d3f291be8-kube-api-access-x7c55\") pod \"cinder-scheduler-0\" (UID: \"63c808d6-07c9-4f88-a262-447d3f291be8\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.209471 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfh7h\" (UniqueName: \"kubernetes.io/projected/b48c437f-bfc5-49c8-bac3-ef4a627fc581-kube-api-access-sfh7h\") pod \"dnsmasq-dns-66996fc899-rbgs7\" (UID: \"b48c437f-bfc5-49c8-bac3-ef4a627fc581\") " pod="openstack/dnsmasq-dns-66996fc899-rbgs7" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.209489 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63c808d6-07c9-4f88-a262-447d3f291be8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"63c808d6-07c9-4f88-a262-447d3f291be8\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.209512 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-config-data\") pod \"cinder-api-0\" (UID: \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\") " pod="openstack/cinder-api-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.209534 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-logs\") pod \"cinder-api-0\" (UID: \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\") " pod="openstack/cinder-api-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.209555 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/63c808d6-07c9-4f88-a262-447d3f291be8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"63c808d6-07c9-4f88-a262-447d3f291be8\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.209571 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-scripts\") pod \"cinder-api-0\" (UID: \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\") " pod="openstack/cinder-api-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.209660 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/63c808d6-07c9-4f88-a262-447d3f291be8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"63c808d6-07c9-4f88-a262-447d3f291be8\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.218160 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63c808d6-07c9-4f88-a262-447d3f291be8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"63c808d6-07c9-4f88-a262-447d3f291be8\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.220229 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/63c808d6-07c9-4f88-a262-447d3f291be8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"63c808d6-07c9-4f88-a262-447d3f291be8\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.220547 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63c808d6-07c9-4f88-a262-447d3f291be8-scripts\") pod \"cinder-scheduler-0\" (UID: \"63c808d6-07c9-4f88-a262-447d3f291be8\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.229697 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63c808d6-07c9-4f88-a262-447d3f291be8-config-data\") pod \"cinder-scheduler-0\" (UID: \"63c808d6-07c9-4f88-a262-447d3f291be8\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.237743 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7c55\" (UniqueName: \"kubernetes.io/projected/63c808d6-07c9-4f88-a262-447d3f291be8-kube-api-access-x7c55\") pod \"cinder-scheduler-0\" (UID: \"63c808d6-07c9-4f88-a262-447d3f291be8\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.311389 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-config-data\") pod \"cinder-api-0\" (UID: \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\") " pod="openstack/cinder-api-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.311453 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-logs\") pod \"cinder-api-0\" (UID: \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\") " pod="openstack/cinder-api-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.311489 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-scripts\") pod \"cinder-api-0\" (UID: \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\") " pod="openstack/cinder-api-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.311552 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b48c437f-bfc5-49c8-bac3-ef4a627fc581-config\") pod \"dnsmasq-dns-66996fc899-rbgs7\" (UID: \"b48c437f-bfc5-49c8-bac3-ef4a627fc581\") " pod="openstack/dnsmasq-dns-66996fc899-rbgs7" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.311577 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b48c437f-bfc5-49c8-bac3-ef4a627fc581-ovsdbserver-sb\") pod \"dnsmasq-dns-66996fc899-rbgs7\" (UID: \"b48c437f-bfc5-49c8-bac3-ef4a627fc581\") " pod="openstack/dnsmasq-dns-66996fc899-rbgs7" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.311605 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-config-data-custom\") pod \"cinder-api-0\" (UID: \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\") " pod="openstack/cinder-api-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.311647 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\") " pod="openstack/cinder-api-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.311669 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b48c437f-bfc5-49c8-bac3-ef4a627fc581-ovsdbserver-nb\") pod \"dnsmasq-dns-66996fc899-rbgs7\" (UID: \"b48c437f-bfc5-49c8-bac3-ef4a627fc581\") " pod="openstack/dnsmasq-dns-66996fc899-rbgs7" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.311705 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\") " pod="openstack/cinder-api-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.311741 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wndnc\" (UniqueName: \"kubernetes.io/projected/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-kube-api-access-wndnc\") pod \"cinder-api-0\" (UID: \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\") " pod="openstack/cinder-api-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.311788 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b48c437f-bfc5-49c8-bac3-ef4a627fc581-dns-svc\") pod \"dnsmasq-dns-66996fc899-rbgs7\" (UID: \"b48c437f-bfc5-49c8-bac3-ef4a627fc581\") " pod="openstack/dnsmasq-dns-66996fc899-rbgs7" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.311837 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfh7h\" (UniqueName: \"kubernetes.io/projected/b48c437f-bfc5-49c8-bac3-ef4a627fc581-kube-api-access-sfh7h\") pod \"dnsmasq-dns-66996fc899-rbgs7\" (UID: \"b48c437f-bfc5-49c8-bac3-ef4a627fc581\") " pod="openstack/dnsmasq-dns-66996fc899-rbgs7" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.312502 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\") " pod="openstack/cinder-api-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.312895 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-logs\") pod \"cinder-api-0\" (UID: \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\") " pod="openstack/cinder-api-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.313282 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b48c437f-bfc5-49c8-bac3-ef4a627fc581-ovsdbserver-sb\") pod \"dnsmasq-dns-66996fc899-rbgs7\" (UID: \"b48c437f-bfc5-49c8-bac3-ef4a627fc581\") " pod="openstack/dnsmasq-dns-66996fc899-rbgs7" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.313869 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b48c437f-bfc5-49c8-bac3-ef4a627fc581-config\") pod \"dnsmasq-dns-66996fc899-rbgs7\" (UID: \"b48c437f-bfc5-49c8-bac3-ef4a627fc581\") " pod="openstack/dnsmasq-dns-66996fc899-rbgs7" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.313875 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b48c437f-bfc5-49c8-bac3-ef4a627fc581-dns-svc\") pod \"dnsmasq-dns-66996fc899-rbgs7\" (UID: \"b48c437f-bfc5-49c8-bac3-ef4a627fc581\") " pod="openstack/dnsmasq-dns-66996fc899-rbgs7" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.316078 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b48c437f-bfc5-49c8-bac3-ef4a627fc581-ovsdbserver-nb\") pod \"dnsmasq-dns-66996fc899-rbgs7\" (UID: \"b48c437f-bfc5-49c8-bac3-ef4a627fc581\") " pod="openstack/dnsmasq-dns-66996fc899-rbgs7" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.317752 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-config-data-custom\") pod \"cinder-api-0\" (UID: \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\") " pod="openstack/cinder-api-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.322591 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\") " pod="openstack/cinder-api-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.334211 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-scripts\") pod \"cinder-api-0\" (UID: \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\") " pod="openstack/cinder-api-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.355989 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-config-data\") pod \"cinder-api-0\" (UID: \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\") " pod="openstack/cinder-api-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.360517 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wndnc\" (UniqueName: \"kubernetes.io/projected/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-kube-api-access-wndnc\") pod \"cinder-api-0\" (UID: \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\") " pod="openstack/cinder-api-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.367582 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfh7h\" (UniqueName: \"kubernetes.io/projected/b48c437f-bfc5-49c8-bac3-ef4a627fc581-kube-api-access-sfh7h\") pod \"dnsmasq-dns-66996fc899-rbgs7\" (UID: \"b48c437f-bfc5-49c8-bac3-ef4a627fc581\") " pod="openstack/dnsmasq-dns-66996fc899-rbgs7" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.410277 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66996fc899-rbgs7"] Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.411060 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66996fc899-rbgs7" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.439824 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bdc9d6cdc-lzb78"] Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.462829 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bdc9d6cdc-lzb78"] Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.462949 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdc9d6cdc-lzb78" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.501043 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.515626 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0789e607-aefa-4c77-9831-050c617faba4-ovsdbserver-nb\") pod \"dnsmasq-dns-7bdc9d6cdc-lzb78\" (UID: \"0789e607-aefa-4c77-9831-050c617faba4\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-lzb78" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.515788 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0789e607-aefa-4c77-9831-050c617faba4-ovsdbserver-sb\") pod \"dnsmasq-dns-7bdc9d6cdc-lzb78\" (UID: \"0789e607-aefa-4c77-9831-050c617faba4\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-lzb78" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.515826 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0789e607-aefa-4c77-9831-050c617faba4-config\") pod \"dnsmasq-dns-7bdc9d6cdc-lzb78\" (UID: \"0789e607-aefa-4c77-9831-050c617faba4\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-lzb78" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.516236 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0789e607-aefa-4c77-9831-050c617faba4-dns-svc\") pod \"dnsmasq-dns-7bdc9d6cdc-lzb78\" (UID: \"0789e607-aefa-4c77-9831-050c617faba4\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-lzb78" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.516323 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrnh5\" (UniqueName: \"kubernetes.io/projected/0789e607-aefa-4c77-9831-050c617faba4-kube-api-access-jrnh5\") pod \"dnsmasq-dns-7bdc9d6cdc-lzb78\" (UID: \"0789e607-aefa-4c77-9831-050c617faba4\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-lzb78" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.521625 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.582353 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7748b8ffbf-c9js9" event={"ID":"d4f04734-09dc-47cf-9395-5f2ca7739a3f","Type":"ContainerStarted","Data":"b7e400d1bfe38afaa2976747f6e7ba3810374d90b750bfff5f80e1b1fff2097d"} Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.583712 4959 generic.go:334] "Generic (PLEG): container finished" podID="541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa" containerID="9e06f9d83c5ab802eb43bda46d60f0f6f1f2cbaf6ba4501be689ff29593d2d57" exitCode=0 Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.583755 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f84c94c5-cbptf" event={"ID":"541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa","Type":"ContainerDied","Data":"9e06f9d83c5ab802eb43bda46d60f0f6f1f2cbaf6ba4501be689ff29593d2d57"} Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.583770 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f84c94c5-cbptf" event={"ID":"541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa","Type":"ContainerStarted","Data":"826338a6065361111c02ba13d3186eaa44cc1ba549ba59fd474468f96344956f"} Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.588205 4959 generic.go:334] "Generic (PLEG): container finished" podID="cd33b52c-3f7f-425e-aa42-7342343067a6" containerID="1fce89d9c71d41fab512a1069c3e290692b1fea1e62f647256e20a6167fe5c1a" exitCode=0 Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.588228 4959 generic.go:334] "Generic (PLEG): container finished" podID="cd33b52c-3f7f-425e-aa42-7342343067a6" containerID="1395f86ab5b523c157ccdd52e29a4d040ff47bbc2a8f66c3696b941b2aa1e1eb" exitCode=2 Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.588258 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd33b52c-3f7f-425e-aa42-7342343067a6","Type":"ContainerDied","Data":"1fce89d9c71d41fab512a1069c3e290692b1fea1e62f647256e20a6167fe5c1a"} Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.588273 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd33b52c-3f7f-425e-aa42-7342343067a6","Type":"ContainerDied","Data":"1395f86ab5b523c157ccdd52e29a4d040ff47bbc2a8f66c3696b941b2aa1e1eb"} Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.590616 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-646c6bbb48-kcmnj" event={"ID":"3a0d41b9-b996-4597-8646-4fb61c6dc354","Type":"ContainerStarted","Data":"ce1f1e59f2f44d13625458ae22ae92973399e2605901eb31ebbdaeb4eec12431"} Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.590642 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-646c6bbb48-kcmnj" event={"ID":"3a0d41b9-b996-4597-8646-4fb61c6dc354","Type":"ContainerStarted","Data":"eacdfcf1036d730fa2b702efb2a8f02cb1290e40489fba042b1f42ba12e22c7b"} Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.591314 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-646c6bbb48-kcmnj" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.591340 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-646c6bbb48-kcmnj" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.592425 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-76c697766b-9qfh8" event={"ID":"a4a56ff7-04cc-48e9-be4a-651a98c06204","Type":"ContainerStarted","Data":"1f59636d543f6960c69b95d0e87b858bfa46f2f1bb6f7de3bfbb0422c1eab754"} Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.596372 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-56c8db7dd-6jq82" event={"ID":"36d7df61-5454-44d6-a7e1-04a0fc4cb621","Type":"ContainerStarted","Data":"546ae5772461e7fefc8f9e277db9a11f529844eba354aa70316106907abbd73c"} Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.596394 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-56c8db7dd-6jq82" event={"ID":"36d7df61-5454-44d6-a7e1-04a0fc4cb621","Type":"ContainerStarted","Data":"8d2b0176f55e84d451a8b517aabfb4ed2d6ce5eb2ae5c69d26bd975d47f40c75"} Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.596414 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-56c8db7dd-6jq82" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.596888 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-56c8db7dd-6jq82" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.618245 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0789e607-aefa-4c77-9831-050c617faba4-config\") pod \"dnsmasq-dns-7bdc9d6cdc-lzb78\" (UID: \"0789e607-aefa-4c77-9831-050c617faba4\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-lzb78" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.618432 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0789e607-aefa-4c77-9831-050c617faba4-dns-svc\") pod \"dnsmasq-dns-7bdc9d6cdc-lzb78\" (UID: \"0789e607-aefa-4c77-9831-050c617faba4\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-lzb78" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.618461 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrnh5\" (UniqueName: \"kubernetes.io/projected/0789e607-aefa-4c77-9831-050c617faba4-kube-api-access-jrnh5\") pod \"dnsmasq-dns-7bdc9d6cdc-lzb78\" (UID: \"0789e607-aefa-4c77-9831-050c617faba4\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-lzb78" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.618511 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0789e607-aefa-4c77-9831-050c617faba4-ovsdbserver-nb\") pod \"dnsmasq-dns-7bdc9d6cdc-lzb78\" (UID: \"0789e607-aefa-4c77-9831-050c617faba4\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-lzb78" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.618597 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0789e607-aefa-4c77-9831-050c617faba4-ovsdbserver-sb\") pod \"dnsmasq-dns-7bdc9d6cdc-lzb78\" (UID: \"0789e607-aefa-4c77-9831-050c617faba4\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-lzb78" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.619515 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0789e607-aefa-4c77-9831-050c617faba4-ovsdbserver-sb\") pod \"dnsmasq-dns-7bdc9d6cdc-lzb78\" (UID: \"0789e607-aefa-4c77-9831-050c617faba4\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-lzb78" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.623230 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0789e607-aefa-4c77-9831-050c617faba4-config\") pod \"dnsmasq-dns-7bdc9d6cdc-lzb78\" (UID: \"0789e607-aefa-4c77-9831-050c617faba4\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-lzb78" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.625588 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0789e607-aefa-4c77-9831-050c617faba4-dns-svc\") pod \"dnsmasq-dns-7bdc9d6cdc-lzb78\" (UID: \"0789e607-aefa-4c77-9831-050c617faba4\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-lzb78" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.631646 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-646c6bbb48-kcmnj" podStartSLOduration=3.631618927 podStartE2EDuration="3.631618927s" podCreationTimestamp="2025-10-07 14:03:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:03:38.631537195 +0000 UTC m=+1100.714941520" watchObservedRunningTime="2025-10-07 14:03:38.631618927 +0000 UTC m=+1100.715023252" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.633214 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0789e607-aefa-4c77-9831-050c617faba4-ovsdbserver-nb\") pod \"dnsmasq-dns-7bdc9d6cdc-lzb78\" (UID: \"0789e607-aefa-4c77-9831-050c617faba4\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-lzb78" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.668130 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrnh5\" (UniqueName: \"kubernetes.io/projected/0789e607-aefa-4c77-9831-050c617faba4-kube-api-access-jrnh5\") pod \"dnsmasq-dns-7bdc9d6cdc-lzb78\" (UID: \"0789e607-aefa-4c77-9831-050c617faba4\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-lzb78" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.674153 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-56c8db7dd-6jq82" podStartSLOduration=6.6741334 podStartE2EDuration="6.6741334s" podCreationTimestamp="2025-10-07 14:03:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:03:38.658689218 +0000 UTC m=+1100.742093543" watchObservedRunningTime="2025-10-07 14:03:38.6741334 +0000 UTC m=+1100.757537725" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.716673 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-d8944fbd8-5mpd2"] Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.718911 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d8944fbd8-5mpd2" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.727270 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.727620 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.727827 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.728061 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-rbsgf" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.748777 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d8944fbd8-5mpd2"] Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.798326 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdc9d6cdc-lzb78" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.842356 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-httpd-config\") pod \"neutron-d8944fbd8-5mpd2\" (UID: \"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a\") " pod="openstack/neutron-d8944fbd8-5mpd2" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.842440 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8kxg\" (UniqueName: \"kubernetes.io/projected/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-kube-api-access-s8kxg\") pod \"neutron-d8944fbd8-5mpd2\" (UID: \"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a\") " pod="openstack/neutron-d8944fbd8-5mpd2" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.842467 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-ovndb-tls-certs\") pod \"neutron-d8944fbd8-5mpd2\" (UID: \"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a\") " pod="openstack/neutron-d8944fbd8-5mpd2" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.842582 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-combined-ca-bundle\") pod \"neutron-d8944fbd8-5mpd2\" (UID: \"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a\") " pod="openstack/neutron-d8944fbd8-5mpd2" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.842614 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-config\") pod \"neutron-d8944fbd8-5mpd2\" (UID: \"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a\") " pod="openstack/neutron-d8944fbd8-5mpd2" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.944993 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8kxg\" (UniqueName: \"kubernetes.io/projected/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-kube-api-access-s8kxg\") pod \"neutron-d8944fbd8-5mpd2\" (UID: \"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a\") " pod="openstack/neutron-d8944fbd8-5mpd2" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.945037 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-ovndb-tls-certs\") pod \"neutron-d8944fbd8-5mpd2\" (UID: \"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a\") " pod="openstack/neutron-d8944fbd8-5mpd2" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.945139 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-combined-ca-bundle\") pod \"neutron-d8944fbd8-5mpd2\" (UID: \"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a\") " pod="openstack/neutron-d8944fbd8-5mpd2" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.945158 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-config\") pod \"neutron-d8944fbd8-5mpd2\" (UID: \"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a\") " pod="openstack/neutron-d8944fbd8-5mpd2" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.945219 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-httpd-config\") pod \"neutron-d8944fbd8-5mpd2\" (UID: \"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a\") " pod="openstack/neutron-d8944fbd8-5mpd2" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.950746 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-combined-ca-bundle\") pod \"neutron-d8944fbd8-5mpd2\" (UID: \"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a\") " pod="openstack/neutron-d8944fbd8-5mpd2" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.952741 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-httpd-config\") pod \"neutron-d8944fbd8-5mpd2\" (UID: \"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a\") " pod="openstack/neutron-d8944fbd8-5mpd2" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.953644 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-config\") pod \"neutron-d8944fbd8-5mpd2\" (UID: \"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a\") " pod="openstack/neutron-d8944fbd8-5mpd2" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.953695 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-ovndb-tls-certs\") pod \"neutron-d8944fbd8-5mpd2\" (UID: \"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a\") " pod="openstack/neutron-d8944fbd8-5mpd2" Oct 07 14:03:38 crc kubenswrapper[4959]: I1007 14:03:38.962834 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8kxg\" (UniqueName: \"kubernetes.io/projected/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-kube-api-access-s8kxg\") pod \"neutron-d8944fbd8-5mpd2\" (UID: \"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a\") " pod="openstack/neutron-d8944fbd8-5mpd2" Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.010691 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66996fc899-rbgs7"] Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.063520 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d8944fbd8-5mpd2" Oct 07 14:03:39 crc kubenswrapper[4959]: W1007 14:03:39.071743 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb48c437f_bfc5_49c8_bac3_ef4a627fc581.slice/crio-1a948a264ff49398f02a60ceb054b3086867c1cae7454824d71ec398b5bfb9ab WatchSource:0}: Error finding container 1a948a264ff49398f02a60ceb054b3086867c1cae7454824d71ec398b5bfb9ab: Status 404 returned error can't find the container with id 1a948a264ff49398f02a60ceb054b3086867c1cae7454824d71ec398b5bfb9ab Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.223062 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.228047 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f84c94c5-cbptf" Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.251773 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 14:03:39 crc kubenswrapper[4959]: W1007 14:03:39.254054 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2636da6a_dd0e_42c1_a6b2_6cd2e37f73f2.slice/crio-eaf9dc2b6f9d85096ebc0bd16f27bec3cea67d3f0c16d4734b39ca9cf93d68e9 WatchSource:0}: Error finding container eaf9dc2b6f9d85096ebc0bd16f27bec3cea67d3f0c16d4734b39ca9cf93d68e9: Status 404 returned error can't find the container with id eaf9dc2b6f9d85096ebc0bd16f27bec3cea67d3f0c16d4734b39ca9cf93d68e9 Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.352857 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-config\") pod \"541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa\" (UID: \"541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa\") " Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.353180 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-ovsdbserver-sb\") pod \"541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa\" (UID: \"541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa\") " Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.353272 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zgcj\" (UniqueName: \"kubernetes.io/projected/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-kube-api-access-9zgcj\") pod \"541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa\" (UID: \"541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa\") " Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.353308 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-ovsdbserver-nb\") pod \"541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa\" (UID: \"541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa\") " Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.353376 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-dns-svc\") pod \"541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa\" (UID: \"541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa\") " Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.365345 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-kube-api-access-9zgcj" (OuterVolumeSpecName: "kube-api-access-9zgcj") pod "541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa" (UID: "541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa"). InnerVolumeSpecName "kube-api-access-9zgcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.380760 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa" (UID: "541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.422205 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa" (UID: "541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.425842 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-config" (OuterVolumeSpecName: "config") pod "541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa" (UID: "541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.446553 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa" (UID: "541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.455840 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-config\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.455876 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.455891 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zgcj\" (UniqueName: \"kubernetes.io/projected/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-kube-api-access-9zgcj\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.455904 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.455917 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.504835 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bdc9d6cdc-lzb78"] Oct 07 14:03:39 crc kubenswrapper[4959]: W1007 14:03:39.568368 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0789e607_aefa_4c77_9831_050c617faba4.slice/crio-f8e6595792d7d30c10249576b5e934172a40ff5551131d27846c7aff68c032f7 WatchSource:0}: Error finding container f8e6595792d7d30c10249576b5e934172a40ff5551131d27846c7aff68c032f7: Status 404 returned error can't find the container with id f8e6595792d7d30c10249576b5e934172a40ff5551131d27846c7aff68c032f7 Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.614309 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f84c94c5-cbptf" event={"ID":"541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa","Type":"ContainerDied","Data":"826338a6065361111c02ba13d3186eaa44cc1ba549ba59fd474468f96344956f"} Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.614358 4959 scope.go:117] "RemoveContainer" containerID="9e06f9d83c5ab802eb43bda46d60f0f6f1f2cbaf6ba4501be689ff29593d2d57" Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.614480 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f84c94c5-cbptf" Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.625279 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"63c808d6-07c9-4f88-a262-447d3f291be8","Type":"ContainerStarted","Data":"df250ebfe1bcccd2e14762da5663acf8e0c62427ec7391d1a45f245147a19940"} Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.628664 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bdc9d6cdc-lzb78" event={"ID":"0789e607-aefa-4c77-9831-050c617faba4","Type":"ContainerStarted","Data":"f8e6595792d7d30c10249576b5e934172a40ff5551131d27846c7aff68c032f7"} Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.632107 4959 generic.go:334] "Generic (PLEG): container finished" podID="36d7df61-5454-44d6-a7e1-04a0fc4cb621" containerID="1b1b17927f49a0930007a18699a586fae07ad24aff87233f1fbe96f25aeaa524" exitCode=1 Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.632155 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-56c8db7dd-6jq82" event={"ID":"36d7df61-5454-44d6-a7e1-04a0fc4cb621","Type":"ContainerDied","Data":"1b1b17927f49a0930007a18699a586fae07ad24aff87233f1fbe96f25aeaa524"} Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.633299 4959 scope.go:117] "RemoveContainer" containerID="1b1b17927f49a0930007a18699a586fae07ad24aff87233f1fbe96f25aeaa524" Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.642341 4959 generic.go:334] "Generic (PLEG): container finished" podID="b48c437f-bfc5-49c8-bac3-ef4a627fc581" containerID="88de4b3787d879a2047c3d2c8303fb1d5519a90725b489e8e9bc0b15becd757a" exitCode=0 Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.642399 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66996fc899-rbgs7" event={"ID":"b48c437f-bfc5-49c8-bac3-ef4a627fc581","Type":"ContainerDied","Data":"88de4b3787d879a2047c3d2c8303fb1d5519a90725b489e8e9bc0b15becd757a"} Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.642424 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66996fc899-rbgs7" event={"ID":"b48c437f-bfc5-49c8-bac3-ef4a627fc581","Type":"ContainerStarted","Data":"1a948a264ff49398f02a60ceb054b3086867c1cae7454824d71ec398b5bfb9ab"} Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.649225 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2","Type":"ContainerStarted","Data":"eaf9dc2b6f9d85096ebc0bd16f27bec3cea67d3f0c16d4734b39ca9cf93d68e9"} Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.727186 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f84c94c5-cbptf"] Oct 07 14:03:39 crc kubenswrapper[4959]: I1007 14:03:39.746863 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f84c94c5-cbptf"] Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.106261 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d8944fbd8-5mpd2"] Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.124659 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5c4946988b-h259p" Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.245750 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.591278 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66996fc899-rbgs7" Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.661499 4959 generic.go:334] "Generic (PLEG): container finished" podID="36d7df61-5454-44d6-a7e1-04a0fc4cb621" containerID="afa2af638aab0c9b79834ec6072fec156d8c03536edc995c4fe9518f565c4710" exitCode=1 Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.662335 4959 scope.go:117] "RemoveContainer" containerID="afa2af638aab0c9b79834ec6072fec156d8c03536edc995c4fe9518f565c4710" Oct 07 14:03:40 crc kubenswrapper[4959]: E1007 14:03:40.662928 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=barbican-api pod=barbican-api-56c8db7dd-6jq82_openstack(36d7df61-5454-44d6-a7e1-04a0fc4cb621)\"" pod="openstack/barbican-api-56c8db7dd-6jq82" podUID="36d7df61-5454-44d6-a7e1-04a0fc4cb621" Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.669979 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66996fc899-rbgs7" Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.672706 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa" path="/var/lib/kubelet/pods/541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa/volumes" Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.686163 4959 generic.go:334] "Generic (PLEG): container finished" podID="0789e607-aefa-4c77-9831-050c617faba4" containerID="5e7ab95dce143648aaba21fda226685f08f198b1d25b25dbe727037e8ee0ee20" exitCode=0 Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.719892 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-56c8db7dd-6jq82" event={"ID":"36d7df61-5454-44d6-a7e1-04a0fc4cb621","Type":"ContainerDied","Data":"afa2af638aab0c9b79834ec6072fec156d8c03536edc995c4fe9518f565c4710"} Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.719945 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66996fc899-rbgs7" event={"ID":"b48c437f-bfc5-49c8-bac3-ef4a627fc581","Type":"ContainerDied","Data":"1a948a264ff49398f02a60ceb054b3086867c1cae7454824d71ec398b5bfb9ab"} Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.719961 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2","Type":"ContainerStarted","Data":"7490c223bd24b944b5ec7d7964daf2d571e39ddb112c4e426907d3484b2e6a7b"} Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.719975 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bdc9d6cdc-lzb78" event={"ID":"0789e607-aefa-4c77-9831-050c617faba4","Type":"ContainerDied","Data":"5e7ab95dce143648aaba21fda226685f08f198b1d25b25dbe727037e8ee0ee20"} Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.719985 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d8944fbd8-5mpd2" event={"ID":"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a","Type":"ContainerStarted","Data":"2568bafb261674b5e4117efa8c6dcc6a1a271e9b1fca93456822da5dfd6f4693"} Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.720003 4959 scope.go:117] "RemoveContainer" containerID="1b1b17927f49a0930007a18699a586fae07ad24aff87233f1fbe96f25aeaa524" Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.793365 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b48c437f-bfc5-49c8-bac3-ef4a627fc581-ovsdbserver-sb\") pod \"b48c437f-bfc5-49c8-bac3-ef4a627fc581\" (UID: \"b48c437f-bfc5-49c8-bac3-ef4a627fc581\") " Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.793578 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b48c437f-bfc5-49c8-bac3-ef4a627fc581-config\") pod \"b48c437f-bfc5-49c8-bac3-ef4a627fc581\" (UID: \"b48c437f-bfc5-49c8-bac3-ef4a627fc581\") " Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.793620 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b48c437f-bfc5-49c8-bac3-ef4a627fc581-ovsdbserver-nb\") pod \"b48c437f-bfc5-49c8-bac3-ef4a627fc581\" (UID: \"b48c437f-bfc5-49c8-bac3-ef4a627fc581\") " Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.793646 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfh7h\" (UniqueName: \"kubernetes.io/projected/b48c437f-bfc5-49c8-bac3-ef4a627fc581-kube-api-access-sfh7h\") pod \"b48c437f-bfc5-49c8-bac3-ef4a627fc581\" (UID: \"b48c437f-bfc5-49c8-bac3-ef4a627fc581\") " Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.793763 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b48c437f-bfc5-49c8-bac3-ef4a627fc581-dns-svc\") pod \"b48c437f-bfc5-49c8-bac3-ef4a627fc581\" (UID: \"b48c437f-bfc5-49c8-bac3-ef4a627fc581\") " Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.799686 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b48c437f-bfc5-49c8-bac3-ef4a627fc581-kube-api-access-sfh7h" (OuterVolumeSpecName: "kube-api-access-sfh7h") pod "b48c437f-bfc5-49c8-bac3-ef4a627fc581" (UID: "b48c437f-bfc5-49c8-bac3-ef4a627fc581"). InnerVolumeSpecName "kube-api-access-sfh7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.817402 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b48c437f-bfc5-49c8-bac3-ef4a627fc581-config" (OuterVolumeSpecName: "config") pod "b48c437f-bfc5-49c8-bac3-ef4a627fc581" (UID: "b48c437f-bfc5-49c8-bac3-ef4a627fc581"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.820259 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b48c437f-bfc5-49c8-bac3-ef4a627fc581-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b48c437f-bfc5-49c8-bac3-ef4a627fc581" (UID: "b48c437f-bfc5-49c8-bac3-ef4a627fc581"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.820708 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b48c437f-bfc5-49c8-bac3-ef4a627fc581-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b48c437f-bfc5-49c8-bac3-ef4a627fc581" (UID: "b48c437f-bfc5-49c8-bac3-ef4a627fc581"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.823086 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b48c437f-bfc5-49c8-bac3-ef4a627fc581-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b48c437f-bfc5-49c8-bac3-ef4a627fc581" (UID: "b48c437f-bfc5-49c8-bac3-ef4a627fc581"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.897129 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b48c437f-bfc5-49c8-bac3-ef4a627fc581-config\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.897158 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b48c437f-bfc5-49c8-bac3-ef4a627fc581-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.897169 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfh7h\" (UniqueName: \"kubernetes.io/projected/b48c437f-bfc5-49c8-bac3-ef4a627fc581-kube-api-access-sfh7h\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.897179 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b48c437f-bfc5-49c8-bac3-ef4a627fc581-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:40 crc kubenswrapper[4959]: I1007 14:03:40.897187 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b48c437f-bfc5-49c8-bac3-ef4a627fc581-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.049704 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66996fc899-rbgs7"] Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.055809 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-66996fc899-rbgs7"] Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.198568 4959 scope.go:117] "RemoveContainer" containerID="88de4b3787d879a2047c3d2c8303fb1d5519a90725b489e8e9bc0b15becd757a" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.415315 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.614607 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd33b52c-3f7f-425e-aa42-7342343067a6-run-httpd\") pod \"cd33b52c-3f7f-425e-aa42-7342343067a6\" (UID: \"cd33b52c-3f7f-425e-aa42-7342343067a6\") " Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.614665 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd33b52c-3f7f-425e-aa42-7342343067a6-config-data\") pod \"cd33b52c-3f7f-425e-aa42-7342343067a6\" (UID: \"cd33b52c-3f7f-425e-aa42-7342343067a6\") " Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.614698 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcw9j\" (UniqueName: \"kubernetes.io/projected/cd33b52c-3f7f-425e-aa42-7342343067a6-kube-api-access-lcw9j\") pod \"cd33b52c-3f7f-425e-aa42-7342343067a6\" (UID: \"cd33b52c-3f7f-425e-aa42-7342343067a6\") " Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.614833 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd33b52c-3f7f-425e-aa42-7342343067a6-combined-ca-bundle\") pod \"cd33b52c-3f7f-425e-aa42-7342343067a6\" (UID: \"cd33b52c-3f7f-425e-aa42-7342343067a6\") " Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.614861 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd33b52c-3f7f-425e-aa42-7342343067a6-log-httpd\") pod \"cd33b52c-3f7f-425e-aa42-7342343067a6\" (UID: \"cd33b52c-3f7f-425e-aa42-7342343067a6\") " Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.614921 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd33b52c-3f7f-425e-aa42-7342343067a6-scripts\") pod \"cd33b52c-3f7f-425e-aa42-7342343067a6\" (UID: \"cd33b52c-3f7f-425e-aa42-7342343067a6\") " Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.614968 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cd33b52c-3f7f-425e-aa42-7342343067a6-sg-core-conf-yaml\") pod \"cd33b52c-3f7f-425e-aa42-7342343067a6\" (UID: \"cd33b52c-3f7f-425e-aa42-7342343067a6\") " Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.617092 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd33b52c-3f7f-425e-aa42-7342343067a6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "cd33b52c-3f7f-425e-aa42-7342343067a6" (UID: "cd33b52c-3f7f-425e-aa42-7342343067a6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.617473 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd33b52c-3f7f-425e-aa42-7342343067a6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "cd33b52c-3f7f-425e-aa42-7342343067a6" (UID: "cd33b52c-3f7f-425e-aa42-7342343067a6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.623302 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd33b52c-3f7f-425e-aa42-7342343067a6-scripts" (OuterVolumeSpecName: "scripts") pod "cd33b52c-3f7f-425e-aa42-7342343067a6" (UID: "cd33b52c-3f7f-425e-aa42-7342343067a6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.623510 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd33b52c-3f7f-425e-aa42-7342343067a6-kube-api-access-lcw9j" (OuterVolumeSpecName: "kube-api-access-lcw9j") pod "cd33b52c-3f7f-425e-aa42-7342343067a6" (UID: "cd33b52c-3f7f-425e-aa42-7342343067a6"). InnerVolumeSpecName "kube-api-access-lcw9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.654282 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd33b52c-3f7f-425e-aa42-7342343067a6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "cd33b52c-3f7f-425e-aa42-7342343067a6" (UID: "cd33b52c-3f7f-425e-aa42-7342343067a6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.696879 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd33b52c-3f7f-425e-aa42-7342343067a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd33b52c-3f7f-425e-aa42-7342343067a6" (UID: "cd33b52c-3f7f-425e-aa42-7342343067a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.716445 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd33b52c-3f7f-425e-aa42-7342343067a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.716470 4959 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd33b52c-3f7f-425e-aa42-7342343067a6-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.716505 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd33b52c-3f7f-425e-aa42-7342343067a6-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.716514 4959 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cd33b52c-3f7f-425e-aa42-7342343067a6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.716522 4959 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd33b52c-3f7f-425e-aa42-7342343067a6-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.716533 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcw9j\" (UniqueName: \"kubernetes.io/projected/cd33b52c-3f7f-425e-aa42-7342343067a6-kube-api-access-lcw9j\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.719283 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd33b52c-3f7f-425e-aa42-7342343067a6-config-data" (OuterVolumeSpecName: "config-data") pod "cd33b52c-3f7f-425e-aa42-7342343067a6" (UID: "cd33b52c-3f7f-425e-aa42-7342343067a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.721807 4959 generic.go:334] "Generic (PLEG): container finished" podID="cd33b52c-3f7f-425e-aa42-7342343067a6" containerID="a7768743c10e8cceb2e9f59d7ad327c68847c768df2193deb859955e0f76a5de" exitCode=0 Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.721868 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd33b52c-3f7f-425e-aa42-7342343067a6","Type":"ContainerDied","Data":"a7768743c10e8cceb2e9f59d7ad327c68847c768df2193deb859955e0f76a5de"} Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.721895 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd33b52c-3f7f-425e-aa42-7342343067a6","Type":"ContainerDied","Data":"4a10abf5a6e5c04f37c12657f3cc13c27886b75c4c9951e8a873277487ae2890"} Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.721911 4959 scope.go:117] "RemoveContainer" containerID="1fce89d9c71d41fab512a1069c3e290692b1fea1e62f647256e20a6167fe5c1a" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.722321 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.730149 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d8944fbd8-5mpd2" event={"ID":"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a","Type":"ContainerStarted","Data":"0db0b06172375d883555d2f1bbb13a63749e39e63b474637ee52f4f21c685469"} Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.730190 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d8944fbd8-5mpd2" event={"ID":"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a","Type":"ContainerStarted","Data":"2da84f09dcd2b60b3aba0ee6f3cf7ede92179d1cd800c821fc83723c68223df4"} Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.731356 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-d8944fbd8-5mpd2" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.734285 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-76c697766b-9qfh8" event={"ID":"a4a56ff7-04cc-48e9-be4a-651a98c06204","Type":"ContainerStarted","Data":"456a6e99b8325fcdf7cae34d018e96214d185ce2fb570eb5deb339588f908676"} Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.734313 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-76c697766b-9qfh8" event={"ID":"a4a56ff7-04cc-48e9-be4a-651a98c06204","Type":"ContainerStarted","Data":"c40406b7daa303815b5750563c1a7f0d24eb2a8667b3513120809608a4d3511f"} Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.738450 4959 scope.go:117] "RemoveContainer" containerID="afa2af638aab0c9b79834ec6072fec156d8c03536edc995c4fe9518f565c4710" Oct 07 14:03:41 crc kubenswrapper[4959]: E1007 14:03:41.738609 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=barbican-api pod=barbican-api-56c8db7dd-6jq82_openstack(36d7df61-5454-44d6-a7e1-04a0fc4cb621)\"" pod="openstack/barbican-api-56c8db7dd-6jq82" podUID="36d7df61-5454-44d6-a7e1-04a0fc4cb621" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.755501 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7748b8ffbf-c9js9" event={"ID":"d4f04734-09dc-47cf-9395-5f2ca7739a3f","Type":"ContainerStarted","Data":"28e60e2af21a4b07e7de2805ecce5d0234666ef0d864d48d7fc84b755dad5fe1"} Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.759276 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-d8944fbd8-5mpd2" podStartSLOduration=3.759256782 podStartE2EDuration="3.759256782s" podCreationTimestamp="2025-10-07 14:03:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:03:41.751202822 +0000 UTC m=+1103.834607147" watchObservedRunningTime="2025-10-07 14:03:41.759256782 +0000 UTC m=+1103.842661107" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.761640 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bdc9d6cdc-lzb78" event={"ID":"0789e607-aefa-4c77-9831-050c617faba4","Type":"ContainerStarted","Data":"f0ebbb837a9273412f30373a4ace9cf1c9ceeeaa6b54fe352f2f2ab03ec94893"} Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.761684 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7bdc9d6cdc-lzb78" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.780231 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-76c697766b-9qfh8" podStartSLOduration=6.584807749 podStartE2EDuration="9.780208151s" podCreationTimestamp="2025-10-07 14:03:32 +0000 UTC" firstStartedPulling="2025-10-07 14:03:37.912231087 +0000 UTC m=+1099.995635412" lastFinishedPulling="2025-10-07 14:03:41.107631489 +0000 UTC m=+1103.191035814" observedRunningTime="2025-10-07 14:03:41.765177358 +0000 UTC m=+1103.848581683" watchObservedRunningTime="2025-10-07 14:03:41.780208151 +0000 UTC m=+1103.863612476" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.802333 4959 scope.go:117] "RemoveContainer" containerID="1395f86ab5b523c157ccdd52e29a4d040ff47bbc2a8f66c3696b941b2aa1e1eb" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.808967 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.818527 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd33b52c-3f7f-425e-aa42-7342343067a6-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.835324 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.877232 4959 scope.go:117] "RemoveContainer" containerID="a7768743c10e8cceb2e9f59d7ad327c68847c768df2193deb859955e0f76a5de" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.887339 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:03:41 crc kubenswrapper[4959]: E1007 14:03:41.887747 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd33b52c-3f7f-425e-aa42-7342343067a6" containerName="sg-core" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.887765 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd33b52c-3f7f-425e-aa42-7342343067a6" containerName="sg-core" Oct 07 14:03:41 crc kubenswrapper[4959]: E1007 14:03:41.887803 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd33b52c-3f7f-425e-aa42-7342343067a6" containerName="proxy-httpd" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.887810 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd33b52c-3f7f-425e-aa42-7342343067a6" containerName="proxy-httpd" Oct 07 14:03:41 crc kubenswrapper[4959]: E1007 14:03:41.887819 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa" containerName="init" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.887825 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa" containerName="init" Oct 07 14:03:41 crc kubenswrapper[4959]: E1007 14:03:41.887838 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b48c437f-bfc5-49c8-bac3-ef4a627fc581" containerName="init" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.887844 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b48c437f-bfc5-49c8-bac3-ef4a627fc581" containerName="init" Oct 07 14:03:41 crc kubenswrapper[4959]: E1007 14:03:41.887854 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd33b52c-3f7f-425e-aa42-7342343067a6" containerName="ceilometer-notification-agent" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.887862 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd33b52c-3f7f-425e-aa42-7342343067a6" containerName="ceilometer-notification-agent" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.888028 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd33b52c-3f7f-425e-aa42-7342343067a6" containerName="ceilometer-notification-agent" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.888038 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="541e2bc0-1e1b-45e2-8d38-2645dcd2a5fa" containerName="init" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.888050 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b48c437f-bfc5-49c8-bac3-ef4a627fc581" containerName="init" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.888064 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd33b52c-3f7f-425e-aa42-7342343067a6" containerName="sg-core" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.888077 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd33b52c-3f7f-425e-aa42-7342343067a6" containerName="proxy-httpd" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.889703 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.889963 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7bdc9d6cdc-lzb78" podStartSLOduration=3.889942247 podStartE2EDuration="3.889942247s" podCreationTimestamp="2025-10-07 14:03:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:03:41.852615213 +0000 UTC m=+1103.936019538" watchObservedRunningTime="2025-10-07 14:03:41.889942247 +0000 UTC m=+1103.973346572" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.892270 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.893042 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.928446 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.938847 4959 scope.go:117] "RemoveContainer" containerID="1fce89d9c71d41fab512a1069c3e290692b1fea1e62f647256e20a6167fe5c1a" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.939686 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7748b8ffbf-c9js9" podStartSLOduration=6.717279859 podStartE2EDuration="9.939672499s" podCreationTimestamp="2025-10-07 14:03:32 +0000 UTC" firstStartedPulling="2025-10-07 14:03:37.91196966 +0000 UTC m=+1099.995373985" lastFinishedPulling="2025-10-07 14:03:41.1343623 +0000 UTC m=+1103.217766625" observedRunningTime="2025-10-07 14:03:41.872527246 +0000 UTC m=+1103.955931571" watchObservedRunningTime="2025-10-07 14:03:41.939672499 +0000 UTC m=+1104.023076824" Oct 07 14:03:41 crc kubenswrapper[4959]: E1007 14:03:41.944484 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fce89d9c71d41fab512a1069c3e290692b1fea1e62f647256e20a6167fe5c1a\": container with ID starting with 1fce89d9c71d41fab512a1069c3e290692b1fea1e62f647256e20a6167fe5c1a not found: ID does not exist" containerID="1fce89d9c71d41fab512a1069c3e290692b1fea1e62f647256e20a6167fe5c1a" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.944528 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fce89d9c71d41fab512a1069c3e290692b1fea1e62f647256e20a6167fe5c1a"} err="failed to get container status \"1fce89d9c71d41fab512a1069c3e290692b1fea1e62f647256e20a6167fe5c1a\": rpc error: code = NotFound desc = could not find container \"1fce89d9c71d41fab512a1069c3e290692b1fea1e62f647256e20a6167fe5c1a\": container with ID starting with 1fce89d9c71d41fab512a1069c3e290692b1fea1e62f647256e20a6167fe5c1a not found: ID does not exist" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.944555 4959 scope.go:117] "RemoveContainer" containerID="1395f86ab5b523c157ccdd52e29a4d040ff47bbc2a8f66c3696b941b2aa1e1eb" Oct 07 14:03:41 crc kubenswrapper[4959]: E1007 14:03:41.951762 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1395f86ab5b523c157ccdd52e29a4d040ff47bbc2a8f66c3696b941b2aa1e1eb\": container with ID starting with 1395f86ab5b523c157ccdd52e29a4d040ff47bbc2a8f66c3696b941b2aa1e1eb not found: ID does not exist" containerID="1395f86ab5b523c157ccdd52e29a4d040ff47bbc2a8f66c3696b941b2aa1e1eb" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.951812 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1395f86ab5b523c157ccdd52e29a4d040ff47bbc2a8f66c3696b941b2aa1e1eb"} err="failed to get container status \"1395f86ab5b523c157ccdd52e29a4d040ff47bbc2a8f66c3696b941b2aa1e1eb\": rpc error: code = NotFound desc = could not find container \"1395f86ab5b523c157ccdd52e29a4d040ff47bbc2a8f66c3696b941b2aa1e1eb\": container with ID starting with 1395f86ab5b523c157ccdd52e29a4d040ff47bbc2a8f66c3696b941b2aa1e1eb not found: ID does not exist" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.951836 4959 scope.go:117] "RemoveContainer" containerID="a7768743c10e8cceb2e9f59d7ad327c68847c768df2193deb859955e0f76a5de" Oct 07 14:03:41 crc kubenswrapper[4959]: E1007 14:03:41.956755 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7768743c10e8cceb2e9f59d7ad327c68847c768df2193deb859955e0f76a5de\": container with ID starting with a7768743c10e8cceb2e9f59d7ad327c68847c768df2193deb859955e0f76a5de not found: ID does not exist" containerID="a7768743c10e8cceb2e9f59d7ad327c68847c768df2193deb859955e0f76a5de" Oct 07 14:03:41 crc kubenswrapper[4959]: I1007 14:03:41.956778 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7768743c10e8cceb2e9f59d7ad327c68847c768df2193deb859955e0f76a5de"} err="failed to get container status \"a7768743c10e8cceb2e9f59d7ad327c68847c768df2193deb859955e0f76a5de\": rpc error: code = NotFound desc = could not find container \"a7768743c10e8cceb2e9f59d7ad327c68847c768df2193deb859955e0f76a5de\": container with ID starting with a7768743c10e8cceb2e9f59d7ad327c68847c768df2193deb859955e0f76a5de not found: ID does not exist" Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.022985 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdf1b525-37e6-46c8-b972-3750fa675cd7-config-data\") pod \"ceilometer-0\" (UID: \"bdf1b525-37e6-46c8-b972-3750fa675cd7\") " pod="openstack/ceilometer-0" Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.023026 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdf1b525-37e6-46c8-b972-3750fa675cd7-scripts\") pod \"ceilometer-0\" (UID: \"bdf1b525-37e6-46c8-b972-3750fa675cd7\") " pod="openstack/ceilometer-0" Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.023695 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdf1b525-37e6-46c8-b972-3750fa675cd7-run-httpd\") pod \"ceilometer-0\" (UID: \"bdf1b525-37e6-46c8-b972-3750fa675cd7\") " pod="openstack/ceilometer-0" Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.023746 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2m9d\" (UniqueName: \"kubernetes.io/projected/bdf1b525-37e6-46c8-b972-3750fa675cd7-kube-api-access-c2m9d\") pod \"ceilometer-0\" (UID: \"bdf1b525-37e6-46c8-b972-3750fa675cd7\") " pod="openstack/ceilometer-0" Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.023774 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bdf1b525-37e6-46c8-b972-3750fa675cd7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bdf1b525-37e6-46c8-b972-3750fa675cd7\") " pod="openstack/ceilometer-0" Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.023839 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdf1b525-37e6-46c8-b972-3750fa675cd7-log-httpd\") pod \"ceilometer-0\" (UID: \"bdf1b525-37e6-46c8-b972-3750fa675cd7\") " pod="openstack/ceilometer-0" Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.023861 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdf1b525-37e6-46c8-b972-3750fa675cd7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bdf1b525-37e6-46c8-b972-3750fa675cd7\") " pod="openstack/ceilometer-0" Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.124931 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdf1b525-37e6-46c8-b972-3750fa675cd7-log-httpd\") pod \"ceilometer-0\" (UID: \"bdf1b525-37e6-46c8-b972-3750fa675cd7\") " pod="openstack/ceilometer-0" Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.125281 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdf1b525-37e6-46c8-b972-3750fa675cd7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bdf1b525-37e6-46c8-b972-3750fa675cd7\") " pod="openstack/ceilometer-0" Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.125326 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdf1b525-37e6-46c8-b972-3750fa675cd7-config-data\") pod \"ceilometer-0\" (UID: \"bdf1b525-37e6-46c8-b972-3750fa675cd7\") " pod="openstack/ceilometer-0" Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.125344 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdf1b525-37e6-46c8-b972-3750fa675cd7-scripts\") pod \"ceilometer-0\" (UID: \"bdf1b525-37e6-46c8-b972-3750fa675cd7\") " pod="openstack/ceilometer-0" Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.125372 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdf1b525-37e6-46c8-b972-3750fa675cd7-log-httpd\") pod \"ceilometer-0\" (UID: \"bdf1b525-37e6-46c8-b972-3750fa675cd7\") " pod="openstack/ceilometer-0" Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.125390 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdf1b525-37e6-46c8-b972-3750fa675cd7-run-httpd\") pod \"ceilometer-0\" (UID: \"bdf1b525-37e6-46c8-b972-3750fa675cd7\") " pod="openstack/ceilometer-0" Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.125466 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2m9d\" (UniqueName: \"kubernetes.io/projected/bdf1b525-37e6-46c8-b972-3750fa675cd7-kube-api-access-c2m9d\") pod \"ceilometer-0\" (UID: \"bdf1b525-37e6-46c8-b972-3750fa675cd7\") " pod="openstack/ceilometer-0" Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.125549 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bdf1b525-37e6-46c8-b972-3750fa675cd7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bdf1b525-37e6-46c8-b972-3750fa675cd7\") " pod="openstack/ceilometer-0" Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.125689 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdf1b525-37e6-46c8-b972-3750fa675cd7-run-httpd\") pod \"ceilometer-0\" (UID: \"bdf1b525-37e6-46c8-b972-3750fa675cd7\") " pod="openstack/ceilometer-0" Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.129761 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdf1b525-37e6-46c8-b972-3750fa675cd7-scripts\") pod \"ceilometer-0\" (UID: \"bdf1b525-37e6-46c8-b972-3750fa675cd7\") " pod="openstack/ceilometer-0" Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.132046 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bdf1b525-37e6-46c8-b972-3750fa675cd7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bdf1b525-37e6-46c8-b972-3750fa675cd7\") " pod="openstack/ceilometer-0" Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.138174 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdf1b525-37e6-46c8-b972-3750fa675cd7-config-data\") pod \"ceilometer-0\" (UID: \"bdf1b525-37e6-46c8-b972-3750fa675cd7\") " pod="openstack/ceilometer-0" Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.138760 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdf1b525-37e6-46c8-b972-3750fa675cd7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bdf1b525-37e6-46c8-b972-3750fa675cd7\") " pod="openstack/ceilometer-0" Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.146575 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2m9d\" (UniqueName: \"kubernetes.io/projected/bdf1b525-37e6-46c8-b972-3750fa675cd7-kube-api-access-c2m9d\") pod \"ceilometer-0\" (UID: \"bdf1b525-37e6-46c8-b972-3750fa675cd7\") " pod="openstack/ceilometer-0" Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.211979 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.291307 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/barbican-api-56c8db7dd-6jq82" Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.674003 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b48c437f-bfc5-49c8-bac3-ef4a627fc581" path="/var/lib/kubelet/pods/b48c437f-bfc5-49c8-bac3-ef4a627fc581/volumes" Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.674734 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd33b52c-3f7f-425e-aa42-7342343067a6" path="/var/lib/kubelet/pods/cd33b52c-3f7f-425e-aa42-7342343067a6/volumes" Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.721416 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:03:42 crc kubenswrapper[4959]: W1007 14:03:42.734531 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbdf1b525_37e6_46c8_b972_3750fa675cd7.slice/crio-f8a8c98d53a271028e1c810511fa6e9b724a84d3a84af6a4ea99228858c6409e WatchSource:0}: Error finding container f8a8c98d53a271028e1c810511fa6e9b724a84d3a84af6a4ea99228858c6409e: Status 404 returned error can't find the container with id f8a8c98d53a271028e1c810511fa6e9b724a84d3a84af6a4ea99228858c6409e Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.790943 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdf1b525-37e6-46c8-b972-3750fa675cd7","Type":"ContainerStarted","Data":"f8a8c98d53a271028e1c810511fa6e9b724a84d3a84af6a4ea99228858c6409e"} Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.798038 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7748b8ffbf-c9js9" event={"ID":"d4f04734-09dc-47cf-9395-5f2ca7739a3f","Type":"ContainerStarted","Data":"cea3707370afc61b0f5312250247fdc2bf13b860f0d2fa4915b204e1d3a8eb6b"} Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.806795 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2","Type":"ContainerStarted","Data":"df3a597df9ffccbb3ad0afa3c4268ef7baa24ccde0d98a77f637f1a39e471bd0"} Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.806952 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2" containerName="cinder-api-log" containerID="cri-o://7490c223bd24b944b5ec7d7964daf2d571e39ddb112c4e426907d3484b2e6a7b" gracePeriod=30 Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.807037 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2" containerName="cinder-api" containerID="cri-o://df3a597df9ffccbb3ad0afa3c4268ef7baa24ccde0d98a77f637f1a39e471bd0" gracePeriod=30 Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.807232 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.813638 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"63c808d6-07c9-4f88-a262-447d3f291be8","Type":"ContainerStarted","Data":"c3d097109d9152828888ff2ef972f260958b2e9e5a05c6a424ab293d15659e10"} Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.816323 4959 scope.go:117] "RemoveContainer" containerID="afa2af638aab0c9b79834ec6072fec156d8c03536edc995c4fe9518f565c4710" Oct 07 14:03:42 crc kubenswrapper[4959]: E1007 14:03:42.816525 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=barbican-api pod=barbican-api-56c8db7dd-6jq82_openstack(36d7df61-5454-44d6-a7e1-04a0fc4cb621)\"" pod="openstack/barbican-api-56c8db7dd-6jq82" podUID="36d7df61-5454-44d6-a7e1-04a0fc4cb621" Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.841328 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.841313152 podStartE2EDuration="4.841313152s" podCreationTimestamp="2025-10-07 14:03:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:03:42.829539 +0000 UTC m=+1104.912943335" watchObservedRunningTime="2025-10-07 14:03:42.841313152 +0000 UTC m=+1104.924717477" Oct 07 14:03:42 crc kubenswrapper[4959]: I1007 14:03:42.857187 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.010986947 podStartE2EDuration="5.857171275s" podCreationTimestamp="2025-10-07 14:03:37 +0000 UTC" firstStartedPulling="2025-10-07 14:03:39.29048463 +0000 UTC m=+1101.373888955" lastFinishedPulling="2025-10-07 14:03:41.136668948 +0000 UTC m=+1103.220073283" observedRunningTime="2025-10-07 14:03:42.855491533 +0000 UTC m=+1104.938895858" watchObservedRunningTime="2025-10-07 14:03:42.857171275 +0000 UTC m=+1104.940575600" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.287373 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-56c8db7dd-6jq82" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.289177 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-56c8db7dd-6jq82" podUID="36d7df61-5454-44d6-a7e1-04a0fc4cb621" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.151:9311/healthcheck\": dial tcp 10.217.0.151:9311: connect: connection refused" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.504749 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.505084 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.666815 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-config-data-custom\") pod \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\" (UID: \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\") " Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.667199 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-scripts\") pod \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\" (UID: \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\") " Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.667362 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-combined-ca-bundle\") pod \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\" (UID: \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\") " Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.667469 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-config-data\") pod \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\" (UID: \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\") " Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.667589 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-logs\") pod \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\" (UID: \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\") " Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.667665 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-etc-machine-id\") pod \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\" (UID: \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\") " Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.667766 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2" (UID: "2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.668174 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wndnc\" (UniqueName: \"kubernetes.io/projected/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-kube-api-access-wndnc\") pod \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\" (UID: \"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2\") " Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.668520 4959 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.677153 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2" (UID: "2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.689403 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-logs" (OuterVolumeSpecName: "logs") pod "2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2" (UID: "2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.691590 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-scripts" (OuterVolumeSpecName: "scripts") pod "2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2" (UID: "2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.696322 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-kube-api-access-wndnc" (OuterVolumeSpecName: "kube-api-access-wndnc") pod "2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2" (UID: "2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2"). InnerVolumeSpecName "kube-api-access-wndnc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.721212 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2" (UID: "2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.750434 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-config-data" (OuterVolumeSpecName: "config-data") pod "2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2" (UID: "2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.773057 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.774116 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.774157 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-logs\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.774167 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wndnc\" (UniqueName: \"kubernetes.io/projected/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-kube-api-access-wndnc\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.774179 4959 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.774195 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.825359 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"63c808d6-07c9-4f88-a262-447d3f291be8","Type":"ContainerStarted","Data":"a4cbe3f833410df29e49c33f863e5f0857f1b27aa3f6e6731133741b56a64af6"} Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.826843 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdf1b525-37e6-46c8-b972-3750fa675cd7","Type":"ContainerStarted","Data":"4eaf793c7493afd3e19b0fce8360a7d6fb26b4001a687df59ddd8cf1af2c649a"} Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.828329 4959 generic.go:334] "Generic (PLEG): container finished" podID="2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2" containerID="df3a597df9ffccbb3ad0afa3c4268ef7baa24ccde0d98a77f637f1a39e471bd0" exitCode=0 Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.828355 4959 generic.go:334] "Generic (PLEG): container finished" podID="2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2" containerID="7490c223bd24b944b5ec7d7964daf2d571e39ddb112c4e426907d3484b2e6a7b" exitCode=143 Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.828382 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.828431 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2","Type":"ContainerDied","Data":"df3a597df9ffccbb3ad0afa3c4268ef7baa24ccde0d98a77f637f1a39e471bd0"} Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.828453 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2","Type":"ContainerDied","Data":"7490c223bd24b944b5ec7d7964daf2d571e39ddb112c4e426907d3484b2e6a7b"} Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.828463 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2","Type":"ContainerDied","Data":"eaf9dc2b6f9d85096ebc0bd16f27bec3cea67d3f0c16d4734b39ca9cf93d68e9"} Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.828478 4959 scope.go:117] "RemoveContainer" containerID="df3a597df9ffccbb3ad0afa3c4268ef7baa24ccde0d98a77f637f1a39e471bd0" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.829005 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-56c8db7dd-6jq82" podUID="36d7df61-5454-44d6-a7e1-04a0fc4cb621" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.151:9311/healthcheck\": dial tcp 10.217.0.151:9311: connect: connection refused" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.829206 4959 scope.go:117] "RemoveContainer" containerID="afa2af638aab0c9b79834ec6072fec156d8c03536edc995c4fe9518f565c4710" Oct 07 14:03:43 crc kubenswrapper[4959]: E1007 14:03:43.829416 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=barbican-api pod=barbican-api-56c8db7dd-6jq82_openstack(36d7df61-5454-44d6-a7e1-04a0fc4cb621)\"" pod="openstack/barbican-api-56c8db7dd-6jq82" podUID="36d7df61-5454-44d6-a7e1-04a0fc4cb621" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.852641 4959 scope.go:117] "RemoveContainer" containerID="7490c223bd24b944b5ec7d7964daf2d571e39ddb112c4e426907d3484b2e6a7b" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.860737 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.867291 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.885583 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 07 14:03:43 crc kubenswrapper[4959]: E1007 14:03:43.886041 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2" containerName="cinder-api-log" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.886133 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2" containerName="cinder-api-log" Oct 07 14:03:43 crc kubenswrapper[4959]: E1007 14:03:43.886219 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2" containerName="cinder-api" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.886288 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2" containerName="cinder-api" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.886480 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2" containerName="cinder-api" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.886537 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2" containerName="cinder-api-log" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.886817 4959 scope.go:117] "RemoveContainer" containerID="df3a597df9ffccbb3ad0afa3c4268ef7baa24ccde0d98a77f637f1a39e471bd0" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.887540 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 07 14:03:43 crc kubenswrapper[4959]: E1007 14:03:43.887676 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df3a597df9ffccbb3ad0afa3c4268ef7baa24ccde0d98a77f637f1a39e471bd0\": container with ID starting with df3a597df9ffccbb3ad0afa3c4268ef7baa24ccde0d98a77f637f1a39e471bd0 not found: ID does not exist" containerID="df3a597df9ffccbb3ad0afa3c4268ef7baa24ccde0d98a77f637f1a39e471bd0" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.887819 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df3a597df9ffccbb3ad0afa3c4268ef7baa24ccde0d98a77f637f1a39e471bd0"} err="failed to get container status \"df3a597df9ffccbb3ad0afa3c4268ef7baa24ccde0d98a77f637f1a39e471bd0\": rpc error: code = NotFound desc = could not find container \"df3a597df9ffccbb3ad0afa3c4268ef7baa24ccde0d98a77f637f1a39e471bd0\": container with ID starting with df3a597df9ffccbb3ad0afa3c4268ef7baa24ccde0d98a77f637f1a39e471bd0 not found: ID does not exist" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.887885 4959 scope.go:117] "RemoveContainer" containerID="7490c223bd24b944b5ec7d7964daf2d571e39ddb112c4e426907d3484b2e6a7b" Oct 07 14:03:43 crc kubenswrapper[4959]: E1007 14:03:43.891324 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7490c223bd24b944b5ec7d7964daf2d571e39ddb112c4e426907d3484b2e6a7b\": container with ID starting with 7490c223bd24b944b5ec7d7964daf2d571e39ddb112c4e426907d3484b2e6a7b not found: ID does not exist" containerID="7490c223bd24b944b5ec7d7964daf2d571e39ddb112c4e426907d3484b2e6a7b" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.891372 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7490c223bd24b944b5ec7d7964daf2d571e39ddb112c4e426907d3484b2e6a7b"} err="failed to get container status \"7490c223bd24b944b5ec7d7964daf2d571e39ddb112c4e426907d3484b2e6a7b\": rpc error: code = NotFound desc = could not find container \"7490c223bd24b944b5ec7d7964daf2d571e39ddb112c4e426907d3484b2e6a7b\": container with ID starting with 7490c223bd24b944b5ec7d7964daf2d571e39ddb112c4e426907d3484b2e6a7b not found: ID does not exist" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.891399 4959 scope.go:117] "RemoveContainer" containerID="df3a597df9ffccbb3ad0afa3c4268ef7baa24ccde0d98a77f637f1a39e471bd0" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.893279 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df3a597df9ffccbb3ad0afa3c4268ef7baa24ccde0d98a77f637f1a39e471bd0"} err="failed to get container status \"df3a597df9ffccbb3ad0afa3c4268ef7baa24ccde0d98a77f637f1a39e471bd0\": rpc error: code = NotFound desc = could not find container \"df3a597df9ffccbb3ad0afa3c4268ef7baa24ccde0d98a77f637f1a39e471bd0\": container with ID starting with df3a597df9ffccbb3ad0afa3c4268ef7baa24ccde0d98a77f637f1a39e471bd0 not found: ID does not exist" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.893315 4959 scope.go:117] "RemoveContainer" containerID="7490c223bd24b944b5ec7d7964daf2d571e39ddb112c4e426907d3484b2e6a7b" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.894198 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.894320 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.894412 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.894756 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7490c223bd24b944b5ec7d7964daf2d571e39ddb112c4e426907d3484b2e6a7b"} err="failed to get container status \"7490c223bd24b944b5ec7d7964daf2d571e39ddb112c4e426907d3484b2e6a7b\": rpc error: code = NotFound desc = could not find container \"7490c223bd24b944b5ec7d7964daf2d571e39ddb112c4e426907d3484b2e6a7b\": container with ID starting with 7490c223bd24b944b5ec7d7964daf2d571e39ddb112c4e426907d3484b2e6a7b not found: ID does not exist" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.910809 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.980918 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/275d4f42-616d-432a-b4b3-932f4df2ab66-scripts\") pod \"cinder-api-0\" (UID: \"275d4f42-616d-432a-b4b3-932f4df2ab66\") " pod="openstack/cinder-api-0" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.981073 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/275d4f42-616d-432a-b4b3-932f4df2ab66-logs\") pod \"cinder-api-0\" (UID: \"275d4f42-616d-432a-b4b3-932f4df2ab66\") " pod="openstack/cinder-api-0" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.981209 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/275d4f42-616d-432a-b4b3-932f4df2ab66-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"275d4f42-616d-432a-b4b3-932f4df2ab66\") " pod="openstack/cinder-api-0" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.981300 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgpnv\" (UniqueName: \"kubernetes.io/projected/275d4f42-616d-432a-b4b3-932f4df2ab66-kube-api-access-rgpnv\") pod \"cinder-api-0\" (UID: \"275d4f42-616d-432a-b4b3-932f4df2ab66\") " pod="openstack/cinder-api-0" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.981559 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/275d4f42-616d-432a-b4b3-932f4df2ab66-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"275d4f42-616d-432a-b4b3-932f4df2ab66\") " pod="openstack/cinder-api-0" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.981704 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/275d4f42-616d-432a-b4b3-932f4df2ab66-etc-machine-id\") pod \"cinder-api-0\" (UID: \"275d4f42-616d-432a-b4b3-932f4df2ab66\") " pod="openstack/cinder-api-0" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.981774 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/275d4f42-616d-432a-b4b3-932f4df2ab66-config-data\") pod \"cinder-api-0\" (UID: \"275d4f42-616d-432a-b4b3-932f4df2ab66\") " pod="openstack/cinder-api-0" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.981976 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/275d4f42-616d-432a-b4b3-932f4df2ab66-public-tls-certs\") pod \"cinder-api-0\" (UID: \"275d4f42-616d-432a-b4b3-932f4df2ab66\") " pod="openstack/cinder-api-0" Oct 07 14:03:43 crc kubenswrapper[4959]: I1007 14:03:43.982047 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/275d4f42-616d-432a-b4b3-932f4df2ab66-config-data-custom\") pod \"cinder-api-0\" (UID: \"275d4f42-616d-432a-b4b3-932f4df2ab66\") " pod="openstack/cinder-api-0" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.083392 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/275d4f42-616d-432a-b4b3-932f4df2ab66-scripts\") pod \"cinder-api-0\" (UID: \"275d4f42-616d-432a-b4b3-932f4df2ab66\") " pod="openstack/cinder-api-0" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.083469 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/275d4f42-616d-432a-b4b3-932f4df2ab66-logs\") pod \"cinder-api-0\" (UID: \"275d4f42-616d-432a-b4b3-932f4df2ab66\") " pod="openstack/cinder-api-0" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.083494 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/275d4f42-616d-432a-b4b3-932f4df2ab66-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"275d4f42-616d-432a-b4b3-932f4df2ab66\") " pod="openstack/cinder-api-0" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.083534 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgpnv\" (UniqueName: \"kubernetes.io/projected/275d4f42-616d-432a-b4b3-932f4df2ab66-kube-api-access-rgpnv\") pod \"cinder-api-0\" (UID: \"275d4f42-616d-432a-b4b3-932f4df2ab66\") " pod="openstack/cinder-api-0" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.083577 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/275d4f42-616d-432a-b4b3-932f4df2ab66-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"275d4f42-616d-432a-b4b3-932f4df2ab66\") " pod="openstack/cinder-api-0" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.083860 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/275d4f42-616d-432a-b4b3-932f4df2ab66-etc-machine-id\") pod \"cinder-api-0\" (UID: \"275d4f42-616d-432a-b4b3-932f4df2ab66\") " pod="openstack/cinder-api-0" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.083948 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/275d4f42-616d-432a-b4b3-932f4df2ab66-config-data\") pod \"cinder-api-0\" (UID: \"275d4f42-616d-432a-b4b3-932f4df2ab66\") " pod="openstack/cinder-api-0" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.084061 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/275d4f42-616d-432a-b4b3-932f4df2ab66-logs\") pod \"cinder-api-0\" (UID: \"275d4f42-616d-432a-b4b3-932f4df2ab66\") " pod="openstack/cinder-api-0" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.084173 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/275d4f42-616d-432a-b4b3-932f4df2ab66-etc-machine-id\") pod \"cinder-api-0\" (UID: \"275d4f42-616d-432a-b4b3-932f4df2ab66\") " pod="openstack/cinder-api-0" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.084262 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/275d4f42-616d-432a-b4b3-932f4df2ab66-config-data-custom\") pod \"cinder-api-0\" (UID: \"275d4f42-616d-432a-b4b3-932f4df2ab66\") " pod="openstack/cinder-api-0" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.084333 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/275d4f42-616d-432a-b4b3-932f4df2ab66-public-tls-certs\") pod \"cinder-api-0\" (UID: \"275d4f42-616d-432a-b4b3-932f4df2ab66\") " pod="openstack/cinder-api-0" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.087531 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/275d4f42-616d-432a-b4b3-932f4df2ab66-scripts\") pod \"cinder-api-0\" (UID: \"275d4f42-616d-432a-b4b3-932f4df2ab66\") " pod="openstack/cinder-api-0" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.089354 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/275d4f42-616d-432a-b4b3-932f4df2ab66-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"275d4f42-616d-432a-b4b3-932f4df2ab66\") " pod="openstack/cinder-api-0" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.090894 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/275d4f42-616d-432a-b4b3-932f4df2ab66-config-data\") pod \"cinder-api-0\" (UID: \"275d4f42-616d-432a-b4b3-932f4df2ab66\") " pod="openstack/cinder-api-0" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.098449 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/275d4f42-616d-432a-b4b3-932f4df2ab66-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"275d4f42-616d-432a-b4b3-932f4df2ab66\") " pod="openstack/cinder-api-0" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.098416 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/275d4f42-616d-432a-b4b3-932f4df2ab66-config-data-custom\") pod \"cinder-api-0\" (UID: \"275d4f42-616d-432a-b4b3-932f4df2ab66\") " pod="openstack/cinder-api-0" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.100654 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgpnv\" (UniqueName: \"kubernetes.io/projected/275d4f42-616d-432a-b4b3-932f4df2ab66-kube-api-access-rgpnv\") pod \"cinder-api-0\" (UID: \"275d4f42-616d-432a-b4b3-932f4df2ab66\") " pod="openstack/cinder-api-0" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.101172 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/275d4f42-616d-432a-b4b3-932f4df2ab66-public-tls-certs\") pod \"cinder-api-0\" (UID: \"275d4f42-616d-432a-b4b3-932f4df2ab66\") " pod="openstack/cinder-api-0" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.214627 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.590696 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-67d6886b4f-lz5gr"] Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.592678 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-67d6886b4f-lz5gr" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.594960 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.595081 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.606093 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-67d6886b4f-lz5gr"] Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.668688 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2" path="/var/lib/kubelet/pods/2636da6a-dd0e-42c1-a6b2-6cd2e37f73f2/volumes" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.695673 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f7cac26-af4a-4db4-b216-f71d41ca9c74-ovndb-tls-certs\") pod \"neutron-67d6886b4f-lz5gr\" (UID: \"8f7cac26-af4a-4db4-b216-f71d41ca9c74\") " pod="openstack/neutron-67d6886b4f-lz5gr" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.695859 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8f7cac26-af4a-4db4-b216-f71d41ca9c74-config\") pod \"neutron-67d6886b4f-lz5gr\" (UID: \"8f7cac26-af4a-4db4-b216-f71d41ca9c74\") " pod="openstack/neutron-67d6886b4f-lz5gr" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.696076 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f7cac26-af4a-4db4-b216-f71d41ca9c74-combined-ca-bundle\") pod \"neutron-67d6886b4f-lz5gr\" (UID: \"8f7cac26-af4a-4db4-b216-f71d41ca9c74\") " pod="openstack/neutron-67d6886b4f-lz5gr" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.696127 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f7cac26-af4a-4db4-b216-f71d41ca9c74-internal-tls-certs\") pod \"neutron-67d6886b4f-lz5gr\" (UID: \"8f7cac26-af4a-4db4-b216-f71d41ca9c74\") " pod="openstack/neutron-67d6886b4f-lz5gr" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.696220 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f7cac26-af4a-4db4-b216-f71d41ca9c74-public-tls-certs\") pod \"neutron-67d6886b4f-lz5gr\" (UID: \"8f7cac26-af4a-4db4-b216-f71d41ca9c74\") " pod="openstack/neutron-67d6886b4f-lz5gr" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.696748 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8f7cac26-af4a-4db4-b216-f71d41ca9c74-httpd-config\") pod \"neutron-67d6886b4f-lz5gr\" (UID: \"8f7cac26-af4a-4db4-b216-f71d41ca9c74\") " pod="openstack/neutron-67d6886b4f-lz5gr" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.696848 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnc25\" (UniqueName: \"kubernetes.io/projected/8f7cac26-af4a-4db4-b216-f71d41ca9c74-kube-api-access-tnc25\") pod \"neutron-67d6886b4f-lz5gr\" (UID: \"8f7cac26-af4a-4db4-b216-f71d41ca9c74\") " pod="openstack/neutron-67d6886b4f-lz5gr" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.713354 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.798203 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f7cac26-af4a-4db4-b216-f71d41ca9c74-combined-ca-bundle\") pod \"neutron-67d6886b4f-lz5gr\" (UID: \"8f7cac26-af4a-4db4-b216-f71d41ca9c74\") " pod="openstack/neutron-67d6886b4f-lz5gr" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.798251 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f7cac26-af4a-4db4-b216-f71d41ca9c74-internal-tls-certs\") pod \"neutron-67d6886b4f-lz5gr\" (UID: \"8f7cac26-af4a-4db4-b216-f71d41ca9c74\") " pod="openstack/neutron-67d6886b4f-lz5gr" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.798285 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f7cac26-af4a-4db4-b216-f71d41ca9c74-public-tls-certs\") pod \"neutron-67d6886b4f-lz5gr\" (UID: \"8f7cac26-af4a-4db4-b216-f71d41ca9c74\") " pod="openstack/neutron-67d6886b4f-lz5gr" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.798333 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8f7cac26-af4a-4db4-b216-f71d41ca9c74-httpd-config\") pod \"neutron-67d6886b4f-lz5gr\" (UID: \"8f7cac26-af4a-4db4-b216-f71d41ca9c74\") " pod="openstack/neutron-67d6886b4f-lz5gr" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.798438 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnc25\" (UniqueName: \"kubernetes.io/projected/8f7cac26-af4a-4db4-b216-f71d41ca9c74-kube-api-access-tnc25\") pod \"neutron-67d6886b4f-lz5gr\" (UID: \"8f7cac26-af4a-4db4-b216-f71d41ca9c74\") " pod="openstack/neutron-67d6886b4f-lz5gr" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.798504 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f7cac26-af4a-4db4-b216-f71d41ca9c74-ovndb-tls-certs\") pod \"neutron-67d6886b4f-lz5gr\" (UID: \"8f7cac26-af4a-4db4-b216-f71d41ca9c74\") " pod="openstack/neutron-67d6886b4f-lz5gr" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.798542 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8f7cac26-af4a-4db4-b216-f71d41ca9c74-config\") pod \"neutron-67d6886b4f-lz5gr\" (UID: \"8f7cac26-af4a-4db4-b216-f71d41ca9c74\") " pod="openstack/neutron-67d6886b4f-lz5gr" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.804035 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8f7cac26-af4a-4db4-b216-f71d41ca9c74-httpd-config\") pod \"neutron-67d6886b4f-lz5gr\" (UID: \"8f7cac26-af4a-4db4-b216-f71d41ca9c74\") " pod="openstack/neutron-67d6886b4f-lz5gr" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.804446 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f7cac26-af4a-4db4-b216-f71d41ca9c74-internal-tls-certs\") pod \"neutron-67d6886b4f-lz5gr\" (UID: \"8f7cac26-af4a-4db4-b216-f71d41ca9c74\") " pod="openstack/neutron-67d6886b4f-lz5gr" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.805050 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8f7cac26-af4a-4db4-b216-f71d41ca9c74-config\") pod \"neutron-67d6886b4f-lz5gr\" (UID: \"8f7cac26-af4a-4db4-b216-f71d41ca9c74\") " pod="openstack/neutron-67d6886b4f-lz5gr" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.805401 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f7cac26-af4a-4db4-b216-f71d41ca9c74-public-tls-certs\") pod \"neutron-67d6886b4f-lz5gr\" (UID: \"8f7cac26-af4a-4db4-b216-f71d41ca9c74\") " pod="openstack/neutron-67d6886b4f-lz5gr" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.806802 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f7cac26-af4a-4db4-b216-f71d41ca9c74-ovndb-tls-certs\") pod \"neutron-67d6886b4f-lz5gr\" (UID: \"8f7cac26-af4a-4db4-b216-f71d41ca9c74\") " pod="openstack/neutron-67d6886b4f-lz5gr" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.807445 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f7cac26-af4a-4db4-b216-f71d41ca9c74-combined-ca-bundle\") pod \"neutron-67d6886b4f-lz5gr\" (UID: \"8f7cac26-af4a-4db4-b216-f71d41ca9c74\") " pod="openstack/neutron-67d6886b4f-lz5gr" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.824197 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnc25\" (UniqueName: \"kubernetes.io/projected/8f7cac26-af4a-4db4-b216-f71d41ca9c74-kube-api-access-tnc25\") pod \"neutron-67d6886b4f-lz5gr\" (UID: \"8f7cac26-af4a-4db4-b216-f71d41ca9c74\") " pod="openstack/neutron-67d6886b4f-lz5gr" Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.839060 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdf1b525-37e6-46c8-b972-3750fa675cd7","Type":"ContainerStarted","Data":"8efefa3d8491a9f20e21845146780f1e5a3dd93f0d529d4c9253d4f128136913"} Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.844749 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"275d4f42-616d-432a-b4b3-932f4df2ab66","Type":"ContainerStarted","Data":"fe6eb6e0305a89e60dba99a930053ca5f4317657dac8516294ab13ef82fee945"} Oct 07 14:03:44 crc kubenswrapper[4959]: I1007 14:03:44.913117 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-67d6886b4f-lz5gr" Oct 07 14:03:45 crc kubenswrapper[4959]: I1007 14:03:45.288481 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-56c8db7dd-6jq82" podUID="36d7df61-5454-44d6-a7e1-04a0fc4cb621" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.151:9311/healthcheck\": dial tcp 10.217.0.151:9311: connect: connection refused" Oct 07 14:03:45 crc kubenswrapper[4959]: I1007 14:03:45.609828 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-67d6886b4f-lz5gr"] Oct 07 14:03:45 crc kubenswrapper[4959]: I1007 14:03:45.853351 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67d6886b4f-lz5gr" event={"ID":"8f7cac26-af4a-4db4-b216-f71d41ca9c74","Type":"ContainerStarted","Data":"3fe471071d432f63b5627daeec6d89f6138928feda803eae474b8be55969280f"} Oct 07 14:03:46 crc kubenswrapper[4959]: I1007 14:03:46.489307 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-744bf8f46d-6twnn" podUID="a588fbdc-fd65-49f6-ab59-4901c7ca690a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Oct 07 14:03:46 crc kubenswrapper[4959]: I1007 14:03:46.612681 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-76d6fc6cbd-2g7sq" podUID="13b27eed-f50f-4474-ace5-0f12e733f6cf" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Oct 07 14:03:47 crc kubenswrapper[4959]: I1007 14:03:47.360979 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-646c6bbb48-kcmnj" Oct 07 14:03:47 crc kubenswrapper[4959]: I1007 14:03:47.388601 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-646c6bbb48-kcmnj" Oct 07 14:03:47 crc kubenswrapper[4959]: I1007 14:03:47.471276 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-56c8db7dd-6jq82"] Oct 07 14:03:47 crc kubenswrapper[4959]: I1007 14:03:47.471465 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-56c8db7dd-6jq82" podUID="36d7df61-5454-44d6-a7e1-04a0fc4cb621" containerName="barbican-api-log" containerID="cri-o://546ae5772461e7fefc8f9e277db9a11f529844eba354aa70316106907abbd73c" gracePeriod=30 Oct 07 14:03:47 crc kubenswrapper[4959]: I1007 14:03:47.471962 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-56c8db7dd-6jq82" podUID="36d7df61-5454-44d6-a7e1-04a0fc4cb621" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.151:9311/healthcheck\": dial tcp 10.217.0.151:9311: connect: connection refused" Oct 07 14:03:47 crc kubenswrapper[4959]: I1007 14:03:47.881912 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-56c8db7dd-6jq82" Oct 07 14:03:47 crc kubenswrapper[4959]: I1007 14:03:47.895000 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"275d4f42-616d-432a-b4b3-932f4df2ab66","Type":"ContainerStarted","Data":"6f946dfbaf935f382727b857152af3e71707f804ae4a41423dd70f9c9706da85"} Oct 07 14:03:47 crc kubenswrapper[4959]: I1007 14:03:47.902303 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67d6886b4f-lz5gr" event={"ID":"8f7cac26-af4a-4db4-b216-f71d41ca9c74","Type":"ContainerStarted","Data":"186a7e86702f892a4f69b97c3dcdda26b061533c2f171c8c1253738c705d112b"} Oct 07 14:03:47 crc kubenswrapper[4959]: I1007 14:03:47.902344 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67d6886b4f-lz5gr" event={"ID":"8f7cac26-af4a-4db4-b216-f71d41ca9c74","Type":"ContainerStarted","Data":"3dc1f32d34a28ac4254fa9f184e206486f1e3cb32c1171ad3f63b0b833904e18"} Oct 07 14:03:47 crc kubenswrapper[4959]: I1007 14:03:47.902387 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-67d6886b4f-lz5gr" Oct 07 14:03:47 crc kubenswrapper[4959]: I1007 14:03:47.918391 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdf1b525-37e6-46c8-b972-3750fa675cd7","Type":"ContainerStarted","Data":"da85e0da09a1be48d799f0a5d54ac72d60b28feed7cfb637414b5d2719b1e0c4"} Oct 07 14:03:47 crc kubenswrapper[4959]: I1007 14:03:47.931292 4959 generic.go:334] "Generic (PLEG): container finished" podID="36d7df61-5454-44d6-a7e1-04a0fc4cb621" containerID="546ae5772461e7fefc8f9e277db9a11f529844eba354aa70316106907abbd73c" exitCode=143 Oct 07 14:03:47 crc kubenswrapper[4959]: I1007 14:03:47.931439 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-56c8db7dd-6jq82" event={"ID":"36d7df61-5454-44d6-a7e1-04a0fc4cb621","Type":"ContainerDied","Data":"546ae5772461e7fefc8f9e277db9a11f529844eba354aa70316106907abbd73c"} Oct 07 14:03:47 crc kubenswrapper[4959]: I1007 14:03:47.931499 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-56c8db7dd-6jq82" event={"ID":"36d7df61-5454-44d6-a7e1-04a0fc4cb621","Type":"ContainerDied","Data":"8d2b0176f55e84d451a8b517aabfb4ed2d6ce5eb2ae5c69d26bd975d47f40c75"} Oct 07 14:03:47 crc kubenswrapper[4959]: I1007 14:03:47.931503 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-56c8db7dd-6jq82" Oct 07 14:03:47 crc kubenswrapper[4959]: I1007 14:03:47.931637 4959 scope.go:117] "RemoveContainer" containerID="afa2af638aab0c9b79834ec6072fec156d8c03536edc995c4fe9518f565c4710" Oct 07 14:03:47 crc kubenswrapper[4959]: I1007 14:03:47.957807 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36d7df61-5454-44d6-a7e1-04a0fc4cb621-config-data\") pod \"36d7df61-5454-44d6-a7e1-04a0fc4cb621\" (UID: \"36d7df61-5454-44d6-a7e1-04a0fc4cb621\") " Oct 07 14:03:47 crc kubenswrapper[4959]: I1007 14:03:47.957862 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/36d7df61-5454-44d6-a7e1-04a0fc4cb621-config-data-custom\") pod \"36d7df61-5454-44d6-a7e1-04a0fc4cb621\" (UID: \"36d7df61-5454-44d6-a7e1-04a0fc4cb621\") " Oct 07 14:03:47 crc kubenswrapper[4959]: I1007 14:03:47.957900 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36d7df61-5454-44d6-a7e1-04a0fc4cb621-logs\") pod \"36d7df61-5454-44d6-a7e1-04a0fc4cb621\" (UID: \"36d7df61-5454-44d6-a7e1-04a0fc4cb621\") " Oct 07 14:03:47 crc kubenswrapper[4959]: I1007 14:03:47.957934 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36d7df61-5454-44d6-a7e1-04a0fc4cb621-combined-ca-bundle\") pod \"36d7df61-5454-44d6-a7e1-04a0fc4cb621\" (UID: \"36d7df61-5454-44d6-a7e1-04a0fc4cb621\") " Oct 07 14:03:47 crc kubenswrapper[4959]: I1007 14:03:47.961245 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-67d6886b4f-lz5gr" podStartSLOduration=3.961224831 podStartE2EDuration="3.961224831s" podCreationTimestamp="2025-10-07 14:03:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:03:47.960050062 +0000 UTC m=+1110.043454407" watchObservedRunningTime="2025-10-07 14:03:47.961224831 +0000 UTC m=+1110.044629156" Oct 07 14:03:47 crc kubenswrapper[4959]: I1007 14:03:47.962216 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36d7df61-5454-44d6-a7e1-04a0fc4cb621-logs" (OuterVolumeSpecName: "logs") pod "36d7df61-5454-44d6-a7e1-04a0fc4cb621" (UID: "36d7df61-5454-44d6-a7e1-04a0fc4cb621"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:03:47 crc kubenswrapper[4959]: I1007 14:03:47.973255 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36d7df61-5454-44d6-a7e1-04a0fc4cb621-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "36d7df61-5454-44d6-a7e1-04a0fc4cb621" (UID: "36d7df61-5454-44d6-a7e1-04a0fc4cb621"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:47 crc kubenswrapper[4959]: I1007 14:03:47.996723 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36d7df61-5454-44d6-a7e1-04a0fc4cb621-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "36d7df61-5454-44d6-a7e1-04a0fc4cb621" (UID: "36d7df61-5454-44d6-a7e1-04a0fc4cb621"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:48 crc kubenswrapper[4959]: I1007 14:03:48.043175 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36d7df61-5454-44d6-a7e1-04a0fc4cb621-config-data" (OuterVolumeSpecName: "config-data") pod "36d7df61-5454-44d6-a7e1-04a0fc4cb621" (UID: "36d7df61-5454-44d6-a7e1-04a0fc4cb621"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:48 crc kubenswrapper[4959]: I1007 14:03:48.059466 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lf98f\" (UniqueName: \"kubernetes.io/projected/36d7df61-5454-44d6-a7e1-04a0fc4cb621-kube-api-access-lf98f\") pod \"36d7df61-5454-44d6-a7e1-04a0fc4cb621\" (UID: \"36d7df61-5454-44d6-a7e1-04a0fc4cb621\") " Oct 07 14:03:48 crc kubenswrapper[4959]: I1007 14:03:48.059954 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36d7df61-5454-44d6-a7e1-04a0fc4cb621-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:48 crc kubenswrapper[4959]: I1007 14:03:48.059972 4959 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/36d7df61-5454-44d6-a7e1-04a0fc4cb621-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:48 crc kubenswrapper[4959]: I1007 14:03:48.059986 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36d7df61-5454-44d6-a7e1-04a0fc4cb621-logs\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:48 crc kubenswrapper[4959]: I1007 14:03:48.059995 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36d7df61-5454-44d6-a7e1-04a0fc4cb621-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:48 crc kubenswrapper[4959]: I1007 14:03:48.065669 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36d7df61-5454-44d6-a7e1-04a0fc4cb621-kube-api-access-lf98f" (OuterVolumeSpecName: "kube-api-access-lf98f") pod "36d7df61-5454-44d6-a7e1-04a0fc4cb621" (UID: "36d7df61-5454-44d6-a7e1-04a0fc4cb621"). InnerVolumeSpecName "kube-api-access-lf98f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:03:48 crc kubenswrapper[4959]: I1007 14:03:48.150482 4959 scope.go:117] "RemoveContainer" containerID="546ae5772461e7fefc8f9e277db9a11f529844eba354aa70316106907abbd73c" Oct 07 14:03:48 crc kubenswrapper[4959]: I1007 14:03:48.165900 4959 scope.go:117] "RemoveContainer" containerID="afa2af638aab0c9b79834ec6072fec156d8c03536edc995c4fe9518f565c4710" Oct 07 14:03:48 crc kubenswrapper[4959]: I1007 14:03:48.167031 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lf98f\" (UniqueName: \"kubernetes.io/projected/36d7df61-5454-44d6-a7e1-04a0fc4cb621-kube-api-access-lf98f\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:48 crc kubenswrapper[4959]: E1007 14:03:48.167131 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afa2af638aab0c9b79834ec6072fec156d8c03536edc995c4fe9518f565c4710\": container with ID starting with afa2af638aab0c9b79834ec6072fec156d8c03536edc995c4fe9518f565c4710 not found: ID does not exist" containerID="afa2af638aab0c9b79834ec6072fec156d8c03536edc995c4fe9518f565c4710" Oct 07 14:03:48 crc kubenswrapper[4959]: I1007 14:03:48.167156 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afa2af638aab0c9b79834ec6072fec156d8c03536edc995c4fe9518f565c4710"} err="failed to get container status \"afa2af638aab0c9b79834ec6072fec156d8c03536edc995c4fe9518f565c4710\": rpc error: code = NotFound desc = could not find container \"afa2af638aab0c9b79834ec6072fec156d8c03536edc995c4fe9518f565c4710\": container with ID starting with afa2af638aab0c9b79834ec6072fec156d8c03536edc995c4fe9518f565c4710 not found: ID does not exist" Oct 07 14:03:48 crc kubenswrapper[4959]: I1007 14:03:48.167178 4959 scope.go:117] "RemoveContainer" containerID="546ae5772461e7fefc8f9e277db9a11f529844eba354aa70316106907abbd73c" Oct 07 14:03:48 crc kubenswrapper[4959]: E1007 14:03:48.172197 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"546ae5772461e7fefc8f9e277db9a11f529844eba354aa70316106907abbd73c\": container with ID starting with 546ae5772461e7fefc8f9e277db9a11f529844eba354aa70316106907abbd73c not found: ID does not exist" containerID="546ae5772461e7fefc8f9e277db9a11f529844eba354aa70316106907abbd73c" Oct 07 14:03:48 crc kubenswrapper[4959]: I1007 14:03:48.172228 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"546ae5772461e7fefc8f9e277db9a11f529844eba354aa70316106907abbd73c"} err="failed to get container status \"546ae5772461e7fefc8f9e277db9a11f529844eba354aa70316106907abbd73c\": rpc error: code = NotFound desc = could not find container \"546ae5772461e7fefc8f9e277db9a11f529844eba354aa70316106907abbd73c\": container with ID starting with 546ae5772461e7fefc8f9e277db9a11f529844eba354aa70316106907abbd73c not found: ID does not exist" Oct 07 14:03:48 crc kubenswrapper[4959]: I1007 14:03:48.269367 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-56c8db7dd-6jq82"] Oct 07 14:03:48 crc kubenswrapper[4959]: I1007 14:03:48.274945 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-56c8db7dd-6jq82"] Oct 07 14:03:48 crc kubenswrapper[4959]: I1007 14:03:48.682896 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36d7df61-5454-44d6-a7e1-04a0fc4cb621" path="/var/lib/kubelet/pods/36d7df61-5454-44d6-a7e1-04a0fc4cb621/volumes" Oct 07 14:03:48 crc kubenswrapper[4959]: I1007 14:03:48.802295 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7bdc9d6cdc-lzb78" Oct 07 14:03:48 crc kubenswrapper[4959]: I1007 14:03:48.811801 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 07 14:03:48 crc kubenswrapper[4959]: I1007 14:03:48.890151 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-748d7644cf-6zpv4"] Oct 07 14:03:48 crc kubenswrapper[4959]: I1007 14:03:48.890720 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-748d7644cf-6zpv4" podUID="0ff263db-2611-4310-a9ee-01e2136c8b1b" containerName="dnsmasq-dns" containerID="cri-o://a5d00af38560f9ea04487166d9f58ac15892c23992b3e960a2f097c414cc9f58" gracePeriod=10 Oct 07 14:03:48 crc kubenswrapper[4959]: I1007 14:03:48.914404 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 14:03:48 crc kubenswrapper[4959]: I1007 14:03:48.961395 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"275d4f42-616d-432a-b4b3-932f4df2ab66","Type":"ContainerStarted","Data":"1aee4832c5fe38b2610f1c97d58358e9a5aaa0b74a8cc6f6627f845f99f928b7"} Oct 07 14:03:48 crc kubenswrapper[4959]: I1007 14:03:48.961830 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 07 14:03:48 crc kubenswrapper[4959]: I1007 14:03:48.969524 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdf1b525-37e6-46c8-b972-3750fa675cd7","Type":"ContainerStarted","Data":"d22923d6cbb6d0255c3d2dc8491f0609142a3434fac743bd79f54081c868bb66"} Oct 07 14:03:48 crc kubenswrapper[4959]: I1007 14:03:48.970043 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 07 14:03:48 crc kubenswrapper[4959]: I1007 14:03:48.977766 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="63c808d6-07c9-4f88-a262-447d3f291be8" containerName="cinder-scheduler" containerID="cri-o://c3d097109d9152828888ff2ef972f260958b2e9e5a05c6a424ab293d15659e10" gracePeriod=30 Oct 07 14:03:48 crc kubenswrapper[4959]: I1007 14:03:48.978039 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="63c808d6-07c9-4f88-a262-447d3f291be8" containerName="probe" containerID="cri-o://a4cbe3f833410df29e49c33f863e5f0857f1b27aa3f6e6731133741b56a64af6" gracePeriod=30 Oct 07 14:03:49 crc kubenswrapper[4959]: I1007 14:03:49.016291 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.016268612 podStartE2EDuration="6.016268612s" podCreationTimestamp="2025-10-07 14:03:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:03:48.999886047 +0000 UTC m=+1111.083290372" watchObservedRunningTime="2025-10-07 14:03:49.016268612 +0000 UTC m=+1111.099672927" Oct 07 14:03:49 crc kubenswrapper[4959]: I1007 14:03:49.047975 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.148820566 podStartE2EDuration="8.047955367s" podCreationTimestamp="2025-10-07 14:03:41 +0000 UTC" firstStartedPulling="2025-10-07 14:03:42.742197438 +0000 UTC m=+1104.825601763" lastFinishedPulling="2025-10-07 14:03:48.641332239 +0000 UTC m=+1110.724736564" observedRunningTime="2025-10-07 14:03:49.036700378 +0000 UTC m=+1111.120104713" watchObservedRunningTime="2025-10-07 14:03:49.047955367 +0000 UTC m=+1111.131359682" Oct 07 14:03:49 crc kubenswrapper[4959]: I1007 14:03:49.454020 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-748d7644cf-6zpv4" Oct 07 14:03:49 crc kubenswrapper[4959]: I1007 14:03:49.534564 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ff263db-2611-4310-a9ee-01e2136c8b1b-config\") pod \"0ff263db-2611-4310-a9ee-01e2136c8b1b\" (UID: \"0ff263db-2611-4310-a9ee-01e2136c8b1b\") " Oct 07 14:03:49 crc kubenswrapper[4959]: I1007 14:03:49.534754 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ff263db-2611-4310-a9ee-01e2136c8b1b-dns-svc\") pod \"0ff263db-2611-4310-a9ee-01e2136c8b1b\" (UID: \"0ff263db-2611-4310-a9ee-01e2136c8b1b\") " Oct 07 14:03:49 crc kubenswrapper[4959]: I1007 14:03:49.534788 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m25xd\" (UniqueName: \"kubernetes.io/projected/0ff263db-2611-4310-a9ee-01e2136c8b1b-kube-api-access-m25xd\") pod \"0ff263db-2611-4310-a9ee-01e2136c8b1b\" (UID: \"0ff263db-2611-4310-a9ee-01e2136c8b1b\") " Oct 07 14:03:49 crc kubenswrapper[4959]: I1007 14:03:49.534848 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ff263db-2611-4310-a9ee-01e2136c8b1b-ovsdbserver-sb\") pod \"0ff263db-2611-4310-a9ee-01e2136c8b1b\" (UID: \"0ff263db-2611-4310-a9ee-01e2136c8b1b\") " Oct 07 14:03:49 crc kubenswrapper[4959]: I1007 14:03:49.534892 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ff263db-2611-4310-a9ee-01e2136c8b1b-ovsdbserver-nb\") pod \"0ff263db-2611-4310-a9ee-01e2136c8b1b\" (UID: \"0ff263db-2611-4310-a9ee-01e2136c8b1b\") " Oct 07 14:03:49 crc kubenswrapper[4959]: I1007 14:03:49.565502 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ff263db-2611-4310-a9ee-01e2136c8b1b-kube-api-access-m25xd" (OuterVolumeSpecName: "kube-api-access-m25xd") pod "0ff263db-2611-4310-a9ee-01e2136c8b1b" (UID: "0ff263db-2611-4310-a9ee-01e2136c8b1b"). InnerVolumeSpecName "kube-api-access-m25xd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:03:49 crc kubenswrapper[4959]: I1007 14:03:49.630826 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ff263db-2611-4310-a9ee-01e2136c8b1b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0ff263db-2611-4310-a9ee-01e2136c8b1b" (UID: "0ff263db-2611-4310-a9ee-01e2136c8b1b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:03:49 crc kubenswrapper[4959]: I1007 14:03:49.639175 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m25xd\" (UniqueName: \"kubernetes.io/projected/0ff263db-2611-4310-a9ee-01e2136c8b1b-kube-api-access-m25xd\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:49 crc kubenswrapper[4959]: I1007 14:03:49.639219 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ff263db-2611-4310-a9ee-01e2136c8b1b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:49 crc kubenswrapper[4959]: I1007 14:03:49.657021 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ff263db-2611-4310-a9ee-01e2136c8b1b-config" (OuterVolumeSpecName: "config") pod "0ff263db-2611-4310-a9ee-01e2136c8b1b" (UID: "0ff263db-2611-4310-a9ee-01e2136c8b1b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:03:49 crc kubenswrapper[4959]: I1007 14:03:49.689749 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ff263db-2611-4310-a9ee-01e2136c8b1b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0ff263db-2611-4310-a9ee-01e2136c8b1b" (UID: "0ff263db-2611-4310-a9ee-01e2136c8b1b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:03:49 crc kubenswrapper[4959]: I1007 14:03:49.714761 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ff263db-2611-4310-a9ee-01e2136c8b1b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0ff263db-2611-4310-a9ee-01e2136c8b1b" (UID: "0ff263db-2611-4310-a9ee-01e2136c8b1b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:03:49 crc kubenswrapper[4959]: I1007 14:03:49.740866 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ff263db-2611-4310-a9ee-01e2136c8b1b-config\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:49 crc kubenswrapper[4959]: I1007 14:03:49.740926 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ff263db-2611-4310-a9ee-01e2136c8b1b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:49 crc kubenswrapper[4959]: I1007 14:03:49.740938 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ff263db-2611-4310-a9ee-01e2136c8b1b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:49 crc kubenswrapper[4959]: I1007 14:03:49.988552 4959 generic.go:334] "Generic (PLEG): container finished" podID="0ff263db-2611-4310-a9ee-01e2136c8b1b" containerID="a5d00af38560f9ea04487166d9f58ac15892c23992b3e960a2f097c414cc9f58" exitCode=0 Oct 07 14:03:49 crc kubenswrapper[4959]: I1007 14:03:49.988627 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-748d7644cf-6zpv4" event={"ID":"0ff263db-2611-4310-a9ee-01e2136c8b1b","Type":"ContainerDied","Data":"a5d00af38560f9ea04487166d9f58ac15892c23992b3e960a2f097c414cc9f58"} Oct 07 14:03:49 crc kubenswrapper[4959]: I1007 14:03:49.988657 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-748d7644cf-6zpv4" event={"ID":"0ff263db-2611-4310-a9ee-01e2136c8b1b","Type":"ContainerDied","Data":"adc28831f6472e2be951a70170e0355baa89eb96c0882c12dc2684466616c423"} Oct 07 14:03:49 crc kubenswrapper[4959]: I1007 14:03:49.988676 4959 scope.go:117] "RemoveContainer" containerID="a5d00af38560f9ea04487166d9f58ac15892c23992b3e960a2f097c414cc9f58" Oct 07 14:03:49 crc kubenswrapper[4959]: I1007 14:03:49.988828 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-748d7644cf-6zpv4" Oct 07 14:03:50 crc kubenswrapper[4959]: I1007 14:03:50.006284 4959 generic.go:334] "Generic (PLEG): container finished" podID="63c808d6-07c9-4f88-a262-447d3f291be8" containerID="a4cbe3f833410df29e49c33f863e5f0857f1b27aa3f6e6731133741b56a64af6" exitCode=0 Oct 07 14:03:50 crc kubenswrapper[4959]: I1007 14:03:50.006380 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"63c808d6-07c9-4f88-a262-447d3f291be8","Type":"ContainerDied","Data":"a4cbe3f833410df29e49c33f863e5f0857f1b27aa3f6e6731133741b56a64af6"} Oct 07 14:03:50 crc kubenswrapper[4959]: I1007 14:03:50.040927 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-748d7644cf-6zpv4"] Oct 07 14:03:50 crc kubenswrapper[4959]: I1007 14:03:50.057456 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-748d7644cf-6zpv4"] Oct 07 14:03:50 crc kubenswrapper[4959]: I1007 14:03:50.063756 4959 scope.go:117] "RemoveContainer" containerID="5096fb136c7910c43dec70435c7f47767bafcf91a92c228f8436148cd995b837" Oct 07 14:03:50 crc kubenswrapper[4959]: I1007 14:03:50.086760 4959 scope.go:117] "RemoveContainer" containerID="a5d00af38560f9ea04487166d9f58ac15892c23992b3e960a2f097c414cc9f58" Oct 07 14:03:50 crc kubenswrapper[4959]: E1007 14:03:50.087288 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5d00af38560f9ea04487166d9f58ac15892c23992b3e960a2f097c414cc9f58\": container with ID starting with a5d00af38560f9ea04487166d9f58ac15892c23992b3e960a2f097c414cc9f58 not found: ID does not exist" containerID="a5d00af38560f9ea04487166d9f58ac15892c23992b3e960a2f097c414cc9f58" Oct 07 14:03:50 crc kubenswrapper[4959]: I1007 14:03:50.087336 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5d00af38560f9ea04487166d9f58ac15892c23992b3e960a2f097c414cc9f58"} err="failed to get container status \"a5d00af38560f9ea04487166d9f58ac15892c23992b3e960a2f097c414cc9f58\": rpc error: code = NotFound desc = could not find container \"a5d00af38560f9ea04487166d9f58ac15892c23992b3e960a2f097c414cc9f58\": container with ID starting with a5d00af38560f9ea04487166d9f58ac15892c23992b3e960a2f097c414cc9f58 not found: ID does not exist" Oct 07 14:03:50 crc kubenswrapper[4959]: I1007 14:03:50.087373 4959 scope.go:117] "RemoveContainer" containerID="5096fb136c7910c43dec70435c7f47767bafcf91a92c228f8436148cd995b837" Oct 07 14:03:50 crc kubenswrapper[4959]: E1007 14:03:50.087742 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5096fb136c7910c43dec70435c7f47767bafcf91a92c228f8436148cd995b837\": container with ID starting with 5096fb136c7910c43dec70435c7f47767bafcf91a92c228f8436148cd995b837 not found: ID does not exist" containerID="5096fb136c7910c43dec70435c7f47767bafcf91a92c228f8436148cd995b837" Oct 07 14:03:50 crc kubenswrapper[4959]: I1007 14:03:50.087762 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5096fb136c7910c43dec70435c7f47767bafcf91a92c228f8436148cd995b837"} err="failed to get container status \"5096fb136c7910c43dec70435c7f47767bafcf91a92c228f8436148cd995b837\": rpc error: code = NotFound desc = could not find container \"5096fb136c7910c43dec70435c7f47767bafcf91a92c228f8436148cd995b837\": container with ID starting with 5096fb136c7910c43dec70435c7f47767bafcf91a92c228f8436148cd995b837 not found: ID does not exist" Oct 07 14:03:50 crc kubenswrapper[4959]: I1007 14:03:50.630339 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:03:50 crc kubenswrapper[4959]: I1007 14:03:50.630395 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:03:50 crc kubenswrapper[4959]: I1007 14:03:50.665478 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ff263db-2611-4310-a9ee-01e2136c8b1b" path="/var/lib/kubelet/pods/0ff263db-2611-4310-a9ee-01e2136c8b1b/volumes" Oct 07 14:03:53 crc kubenswrapper[4959]: I1007 14:03:53.428703 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 07 14:03:53 crc kubenswrapper[4959]: I1007 14:03:53.515225 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7c55\" (UniqueName: \"kubernetes.io/projected/63c808d6-07c9-4f88-a262-447d3f291be8-kube-api-access-x7c55\") pod \"63c808d6-07c9-4f88-a262-447d3f291be8\" (UID: \"63c808d6-07c9-4f88-a262-447d3f291be8\") " Oct 07 14:03:53 crc kubenswrapper[4959]: I1007 14:03:53.515372 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63c808d6-07c9-4f88-a262-447d3f291be8-config-data\") pod \"63c808d6-07c9-4f88-a262-447d3f291be8\" (UID: \"63c808d6-07c9-4f88-a262-447d3f291be8\") " Oct 07 14:03:53 crc kubenswrapper[4959]: I1007 14:03:53.515458 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63c808d6-07c9-4f88-a262-447d3f291be8-scripts\") pod \"63c808d6-07c9-4f88-a262-447d3f291be8\" (UID: \"63c808d6-07c9-4f88-a262-447d3f291be8\") " Oct 07 14:03:53 crc kubenswrapper[4959]: I1007 14:03:53.515474 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/63c808d6-07c9-4f88-a262-447d3f291be8-etc-machine-id\") pod \"63c808d6-07c9-4f88-a262-447d3f291be8\" (UID: \"63c808d6-07c9-4f88-a262-447d3f291be8\") " Oct 07 14:03:53 crc kubenswrapper[4959]: I1007 14:03:53.515495 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/63c808d6-07c9-4f88-a262-447d3f291be8-config-data-custom\") pod \"63c808d6-07c9-4f88-a262-447d3f291be8\" (UID: \"63c808d6-07c9-4f88-a262-447d3f291be8\") " Oct 07 14:03:53 crc kubenswrapper[4959]: I1007 14:03:53.515524 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63c808d6-07c9-4f88-a262-447d3f291be8-combined-ca-bundle\") pod \"63c808d6-07c9-4f88-a262-447d3f291be8\" (UID: \"63c808d6-07c9-4f88-a262-447d3f291be8\") " Oct 07 14:03:53 crc kubenswrapper[4959]: I1007 14:03:53.516363 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/63c808d6-07c9-4f88-a262-447d3f291be8-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "63c808d6-07c9-4f88-a262-447d3f291be8" (UID: "63c808d6-07c9-4f88-a262-447d3f291be8"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 14:03:53 crc kubenswrapper[4959]: I1007 14:03:53.521353 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63c808d6-07c9-4f88-a262-447d3f291be8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "63c808d6-07c9-4f88-a262-447d3f291be8" (UID: "63c808d6-07c9-4f88-a262-447d3f291be8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:53 crc kubenswrapper[4959]: I1007 14:03:53.521465 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63c808d6-07c9-4f88-a262-447d3f291be8-scripts" (OuterVolumeSpecName: "scripts") pod "63c808d6-07c9-4f88-a262-447d3f291be8" (UID: "63c808d6-07c9-4f88-a262-447d3f291be8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:53 crc kubenswrapper[4959]: I1007 14:03:53.526783 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63c808d6-07c9-4f88-a262-447d3f291be8-kube-api-access-x7c55" (OuterVolumeSpecName: "kube-api-access-x7c55") pod "63c808d6-07c9-4f88-a262-447d3f291be8" (UID: "63c808d6-07c9-4f88-a262-447d3f291be8"). InnerVolumeSpecName "kube-api-access-x7c55". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:03:53 crc kubenswrapper[4959]: I1007 14:03:53.572116 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63c808d6-07c9-4f88-a262-447d3f291be8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "63c808d6-07c9-4f88-a262-447d3f291be8" (UID: "63c808d6-07c9-4f88-a262-447d3f291be8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:53 crc kubenswrapper[4959]: I1007 14:03:53.617345 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63c808d6-07c9-4f88-a262-447d3f291be8-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:53 crc kubenswrapper[4959]: I1007 14:03:53.617426 4959 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/63c808d6-07c9-4f88-a262-447d3f291be8-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:53 crc kubenswrapper[4959]: I1007 14:03:53.617438 4959 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/63c808d6-07c9-4f88-a262-447d3f291be8-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:53 crc kubenswrapper[4959]: I1007 14:03:53.617446 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63c808d6-07c9-4f88-a262-447d3f291be8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:53 crc kubenswrapper[4959]: I1007 14:03:53.617476 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7c55\" (UniqueName: \"kubernetes.io/projected/63c808d6-07c9-4f88-a262-447d3f291be8-kube-api-access-x7c55\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:53 crc kubenswrapper[4959]: I1007 14:03:53.635457 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63c808d6-07c9-4f88-a262-447d3f291be8-config-data" (OuterVolumeSpecName: "config-data") pod "63c808d6-07c9-4f88-a262-447d3f291be8" (UID: "63c808d6-07c9-4f88-a262-447d3f291be8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:03:53 crc kubenswrapper[4959]: I1007 14:03:53.718650 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63c808d6-07c9-4f88-a262-447d3f291be8-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.044834 4959 generic.go:334] "Generic (PLEG): container finished" podID="63c808d6-07c9-4f88-a262-447d3f291be8" containerID="c3d097109d9152828888ff2ef972f260958b2e9e5a05c6a424ab293d15659e10" exitCode=0 Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.044896 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"63c808d6-07c9-4f88-a262-447d3f291be8","Type":"ContainerDied","Data":"c3d097109d9152828888ff2ef972f260958b2e9e5a05c6a424ab293d15659e10"} Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.044944 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"63c808d6-07c9-4f88-a262-447d3f291be8","Type":"ContainerDied","Data":"df250ebfe1bcccd2e14762da5663acf8e0c62427ec7391d1a45f245147a19940"} Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.044965 4959 scope.go:117] "RemoveContainer" containerID="a4cbe3f833410df29e49c33f863e5f0857f1b27aa3f6e6731133741b56a64af6" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.044968 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.078032 4959 scope.go:117] "RemoveContainer" containerID="c3d097109d9152828888ff2ef972f260958b2e9e5a05c6a424ab293d15659e10" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.111830 4959 scope.go:117] "RemoveContainer" containerID="a4cbe3f833410df29e49c33f863e5f0857f1b27aa3f6e6731133741b56a64af6" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.112249 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 14:03:54 crc kubenswrapper[4959]: E1007 14:03:54.112404 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4cbe3f833410df29e49c33f863e5f0857f1b27aa3f6e6731133741b56a64af6\": container with ID starting with a4cbe3f833410df29e49c33f863e5f0857f1b27aa3f6e6731133741b56a64af6 not found: ID does not exist" containerID="a4cbe3f833410df29e49c33f863e5f0857f1b27aa3f6e6731133741b56a64af6" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.112433 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4cbe3f833410df29e49c33f863e5f0857f1b27aa3f6e6731133741b56a64af6"} err="failed to get container status \"a4cbe3f833410df29e49c33f863e5f0857f1b27aa3f6e6731133741b56a64af6\": rpc error: code = NotFound desc = could not find container \"a4cbe3f833410df29e49c33f863e5f0857f1b27aa3f6e6731133741b56a64af6\": container with ID starting with a4cbe3f833410df29e49c33f863e5f0857f1b27aa3f6e6731133741b56a64af6 not found: ID does not exist" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.112455 4959 scope.go:117] "RemoveContainer" containerID="c3d097109d9152828888ff2ef972f260958b2e9e5a05c6a424ab293d15659e10" Oct 07 14:03:54 crc kubenswrapper[4959]: E1007 14:03:54.112763 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3d097109d9152828888ff2ef972f260958b2e9e5a05c6a424ab293d15659e10\": container with ID starting with c3d097109d9152828888ff2ef972f260958b2e9e5a05c6a424ab293d15659e10 not found: ID does not exist" containerID="c3d097109d9152828888ff2ef972f260958b2e9e5a05c6a424ab293d15659e10" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.112792 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3d097109d9152828888ff2ef972f260958b2e9e5a05c6a424ab293d15659e10"} err="failed to get container status \"c3d097109d9152828888ff2ef972f260958b2e9e5a05c6a424ab293d15659e10\": rpc error: code = NotFound desc = could not find container \"c3d097109d9152828888ff2ef972f260958b2e9e5a05c6a424ab293d15659e10\": container with ID starting with c3d097109d9152828888ff2ef972f260958b2e9e5a05c6a424ab293d15659e10 not found: ID does not exist" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.125803 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.139069 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 14:03:54 crc kubenswrapper[4959]: E1007 14:03:54.139497 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ff263db-2611-4310-a9ee-01e2136c8b1b" containerName="init" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.139516 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ff263db-2611-4310-a9ee-01e2136c8b1b" containerName="init" Oct 07 14:03:54 crc kubenswrapper[4959]: E1007 14:03:54.139533 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36d7df61-5454-44d6-a7e1-04a0fc4cb621" containerName="barbican-api-log" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.139541 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="36d7df61-5454-44d6-a7e1-04a0fc4cb621" containerName="barbican-api-log" Oct 07 14:03:54 crc kubenswrapper[4959]: E1007 14:03:54.139550 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63c808d6-07c9-4f88-a262-447d3f291be8" containerName="probe" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.139558 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="63c808d6-07c9-4f88-a262-447d3f291be8" containerName="probe" Oct 07 14:03:54 crc kubenswrapper[4959]: E1007 14:03:54.139571 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63c808d6-07c9-4f88-a262-447d3f291be8" containerName="cinder-scheduler" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.139576 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="63c808d6-07c9-4f88-a262-447d3f291be8" containerName="cinder-scheduler" Oct 07 14:03:54 crc kubenswrapper[4959]: E1007 14:03:54.139586 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36d7df61-5454-44d6-a7e1-04a0fc4cb621" containerName="barbican-api" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.139592 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="36d7df61-5454-44d6-a7e1-04a0fc4cb621" containerName="barbican-api" Oct 07 14:03:54 crc kubenswrapper[4959]: E1007 14:03:54.139600 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ff263db-2611-4310-a9ee-01e2136c8b1b" containerName="dnsmasq-dns" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.139605 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ff263db-2611-4310-a9ee-01e2136c8b1b" containerName="dnsmasq-dns" Oct 07 14:03:54 crc kubenswrapper[4959]: E1007 14:03:54.139619 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36d7df61-5454-44d6-a7e1-04a0fc4cb621" containerName="barbican-api" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.139626 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="36d7df61-5454-44d6-a7e1-04a0fc4cb621" containerName="barbican-api" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.139797 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="63c808d6-07c9-4f88-a262-447d3f291be8" containerName="probe" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.139805 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="36d7df61-5454-44d6-a7e1-04a0fc4cb621" containerName="barbican-api" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.139816 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="63c808d6-07c9-4f88-a262-447d3f291be8" containerName="cinder-scheduler" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.139834 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ff263db-2611-4310-a9ee-01e2136c8b1b" containerName="dnsmasq-dns" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.139841 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="36d7df61-5454-44d6-a7e1-04a0fc4cb621" containerName="barbican-api-log" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.140165 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="36d7df61-5454-44d6-a7e1-04a0fc4cb621" containerName="barbican-api" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.140799 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.140935 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.143651 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.226592 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/126adfe4-0ab1-4952-9a03-526fb74cad41-config-data\") pod \"cinder-scheduler-0\" (UID: \"126adfe4-0ab1-4952-9a03-526fb74cad41\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.226637 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/126adfe4-0ab1-4952-9a03-526fb74cad41-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"126adfe4-0ab1-4952-9a03-526fb74cad41\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.226658 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xsqx\" (UniqueName: \"kubernetes.io/projected/126adfe4-0ab1-4952-9a03-526fb74cad41-kube-api-access-7xsqx\") pod \"cinder-scheduler-0\" (UID: \"126adfe4-0ab1-4952-9a03-526fb74cad41\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.226791 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/126adfe4-0ab1-4952-9a03-526fb74cad41-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"126adfe4-0ab1-4952-9a03-526fb74cad41\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.226821 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/126adfe4-0ab1-4952-9a03-526fb74cad41-scripts\") pod \"cinder-scheduler-0\" (UID: \"126adfe4-0ab1-4952-9a03-526fb74cad41\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.226845 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/126adfe4-0ab1-4952-9a03-526fb74cad41-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"126adfe4-0ab1-4952-9a03-526fb74cad41\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.329136 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/126adfe4-0ab1-4952-9a03-526fb74cad41-config-data\") pod \"cinder-scheduler-0\" (UID: \"126adfe4-0ab1-4952-9a03-526fb74cad41\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.329467 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/126adfe4-0ab1-4952-9a03-526fb74cad41-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"126adfe4-0ab1-4952-9a03-526fb74cad41\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.329487 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xsqx\" (UniqueName: \"kubernetes.io/projected/126adfe4-0ab1-4952-9a03-526fb74cad41-kube-api-access-7xsqx\") pod \"cinder-scheduler-0\" (UID: \"126adfe4-0ab1-4952-9a03-526fb74cad41\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.329579 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/126adfe4-0ab1-4952-9a03-526fb74cad41-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"126adfe4-0ab1-4952-9a03-526fb74cad41\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.329596 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/126adfe4-0ab1-4952-9a03-526fb74cad41-scripts\") pod \"cinder-scheduler-0\" (UID: \"126adfe4-0ab1-4952-9a03-526fb74cad41\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.330152 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/126adfe4-0ab1-4952-9a03-526fb74cad41-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"126adfe4-0ab1-4952-9a03-526fb74cad41\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.330094 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/126adfe4-0ab1-4952-9a03-526fb74cad41-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"126adfe4-0ab1-4952-9a03-526fb74cad41\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.334579 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/126adfe4-0ab1-4952-9a03-526fb74cad41-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"126adfe4-0ab1-4952-9a03-526fb74cad41\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.335422 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/126adfe4-0ab1-4952-9a03-526fb74cad41-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"126adfe4-0ab1-4952-9a03-526fb74cad41\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.337536 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/126adfe4-0ab1-4952-9a03-526fb74cad41-scripts\") pod \"cinder-scheduler-0\" (UID: \"126adfe4-0ab1-4952-9a03-526fb74cad41\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.340535 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/126adfe4-0ab1-4952-9a03-526fb74cad41-config-data\") pod \"cinder-scheduler-0\" (UID: \"126adfe4-0ab1-4952-9a03-526fb74cad41\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.349837 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xsqx\" (UniqueName: \"kubernetes.io/projected/126adfe4-0ab1-4952-9a03-526fb74cad41-kube-api-access-7xsqx\") pod \"cinder-scheduler-0\" (UID: \"126adfe4-0ab1-4952-9a03-526fb74cad41\") " pod="openstack/cinder-scheduler-0" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.468026 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.674092 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63c808d6-07c9-4f88-a262-447d3f291be8" path="/var/lib/kubelet/pods/63c808d6-07c9-4f88-a262-447d3f291be8/volumes" Oct 07 14:03:54 crc kubenswrapper[4959]: I1007 14:03:54.926962 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 07 14:03:55 crc kubenswrapper[4959]: I1007 14:03:55.053617 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"126adfe4-0ab1-4952-9a03-526fb74cad41","Type":"ContainerStarted","Data":"031c2409dfee316ca62ce32b80b411b5c56bf7da136006b4f5ae4db9b36b1898"} Oct 07 14:03:56 crc kubenswrapper[4959]: I1007 14:03:56.075611 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"126adfe4-0ab1-4952-9a03-526fb74cad41","Type":"ContainerStarted","Data":"c43fc4c2ac171b945cec8977d99d0c809ad5926b1d56f67cd846890fc51d183a"} Oct 07 14:03:56 crc kubenswrapper[4959]: I1007 14:03:56.076248 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"126adfe4-0ab1-4952-9a03-526fb74cad41","Type":"ContainerStarted","Data":"cea5bca4910bcd9f853447773f5e89f815d0ad84a240a853f75ae8507f9c940e"} Oct 07 14:03:56 crc kubenswrapper[4959]: I1007 14:03:56.103078 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.103062577 podStartE2EDuration="2.103062577s" podCreationTimestamp="2025-10-07 14:03:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:03:56.098694269 +0000 UTC m=+1118.182098594" watchObservedRunningTime="2025-10-07 14:03:56.103062577 +0000 UTC m=+1118.186466902" Oct 07 14:03:56 crc kubenswrapper[4959]: I1007 14:03:56.126411 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 07 14:03:58 crc kubenswrapper[4959]: I1007 14:03:58.314319 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:03:58 crc kubenswrapper[4959]: I1007 14:03:58.526965 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-76d6fc6cbd-2g7sq" Oct 07 14:03:59 crc kubenswrapper[4959]: I1007 14:03:59.468906 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 07 14:04:00 crc kubenswrapper[4959]: I1007 14:04:00.023740 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:04:00 crc kubenswrapper[4959]: I1007 14:04:00.268045 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-76d6fc6cbd-2g7sq" Oct 07 14:04:00 crc kubenswrapper[4959]: I1007 14:04:00.329733 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-744bf8f46d-6twnn"] Oct 07 14:04:00 crc kubenswrapper[4959]: I1007 14:04:00.329990 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-744bf8f46d-6twnn" podUID="a588fbdc-fd65-49f6-ab59-4901c7ca690a" containerName="horizon-log" containerID="cri-o://0b8d35aaa990504cd841339a7496639b4a5b31cf2af6d3b62946d1c9d29fa02f" gracePeriod=30 Oct 07 14:04:00 crc kubenswrapper[4959]: I1007 14:04:00.330133 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-744bf8f46d-6twnn" podUID="a588fbdc-fd65-49f6-ab59-4901c7ca690a" containerName="horizon" containerID="cri-o://76d664db4d06017fdcd093f711c6c672a5feacc3d610e94a71a01f0793720247" gracePeriod=30 Oct 07 14:04:01 crc kubenswrapper[4959]: I1007 14:04:01.011207 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5c4946988b-h259p" Oct 07 14:04:01 crc kubenswrapper[4959]: I1007 14:04:01.473493 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5c77f969b5-4w2xs" Oct 07 14:04:03 crc kubenswrapper[4959]: I1007 14:04:03.069253 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 07 14:04:03 crc kubenswrapper[4959]: I1007 14:04:03.070247 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 07 14:04:03 crc kubenswrapper[4959]: I1007 14:04:03.073924 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 07 14:04:03 crc kubenswrapper[4959]: I1007 14:04:03.076029 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-kbp52" Oct 07 14:04:03 crc kubenswrapper[4959]: I1007 14:04:03.076154 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 07 14:04:03 crc kubenswrapper[4959]: I1007 14:04:03.081401 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 07 14:04:03 crc kubenswrapper[4959]: I1007 14:04:03.192567 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9ba3dd4a-23e6-4617-ae86-d0cc64bec9d8-openstack-config\") pod \"openstackclient\" (UID: \"9ba3dd4a-23e6-4617-ae86-d0cc64bec9d8\") " pod="openstack/openstackclient" Oct 07 14:04:03 crc kubenswrapper[4959]: I1007 14:04:03.192959 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ba3dd4a-23e6-4617-ae86-d0cc64bec9d8-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9ba3dd4a-23e6-4617-ae86-d0cc64bec9d8\") " pod="openstack/openstackclient" Oct 07 14:04:03 crc kubenswrapper[4959]: I1007 14:04:03.193226 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4g8s\" (UniqueName: \"kubernetes.io/projected/9ba3dd4a-23e6-4617-ae86-d0cc64bec9d8-kube-api-access-s4g8s\") pod \"openstackclient\" (UID: \"9ba3dd4a-23e6-4617-ae86-d0cc64bec9d8\") " pod="openstack/openstackclient" Oct 07 14:04:03 crc kubenswrapper[4959]: I1007 14:04:03.193358 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9ba3dd4a-23e6-4617-ae86-d0cc64bec9d8-openstack-config-secret\") pod \"openstackclient\" (UID: \"9ba3dd4a-23e6-4617-ae86-d0cc64bec9d8\") " pod="openstack/openstackclient" Oct 07 14:04:03 crc kubenswrapper[4959]: I1007 14:04:03.294872 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9ba3dd4a-23e6-4617-ae86-d0cc64bec9d8-openstack-config-secret\") pod \"openstackclient\" (UID: \"9ba3dd4a-23e6-4617-ae86-d0cc64bec9d8\") " pod="openstack/openstackclient" Oct 07 14:04:03 crc kubenswrapper[4959]: I1007 14:04:03.295000 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9ba3dd4a-23e6-4617-ae86-d0cc64bec9d8-openstack-config\") pod \"openstackclient\" (UID: \"9ba3dd4a-23e6-4617-ae86-d0cc64bec9d8\") " pod="openstack/openstackclient" Oct 07 14:04:03 crc kubenswrapper[4959]: I1007 14:04:03.295057 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ba3dd4a-23e6-4617-ae86-d0cc64bec9d8-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9ba3dd4a-23e6-4617-ae86-d0cc64bec9d8\") " pod="openstack/openstackclient" Oct 07 14:04:03 crc kubenswrapper[4959]: I1007 14:04:03.295141 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4g8s\" (UniqueName: \"kubernetes.io/projected/9ba3dd4a-23e6-4617-ae86-d0cc64bec9d8-kube-api-access-s4g8s\") pod \"openstackclient\" (UID: \"9ba3dd4a-23e6-4617-ae86-d0cc64bec9d8\") " pod="openstack/openstackclient" Oct 07 14:04:03 crc kubenswrapper[4959]: I1007 14:04:03.295913 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9ba3dd4a-23e6-4617-ae86-d0cc64bec9d8-openstack-config\") pod \"openstackclient\" (UID: \"9ba3dd4a-23e6-4617-ae86-d0cc64bec9d8\") " pod="openstack/openstackclient" Oct 07 14:04:03 crc kubenswrapper[4959]: I1007 14:04:03.300680 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9ba3dd4a-23e6-4617-ae86-d0cc64bec9d8-openstack-config-secret\") pod \"openstackclient\" (UID: \"9ba3dd4a-23e6-4617-ae86-d0cc64bec9d8\") " pod="openstack/openstackclient" Oct 07 14:04:03 crc kubenswrapper[4959]: I1007 14:04:03.300911 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ba3dd4a-23e6-4617-ae86-d0cc64bec9d8-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9ba3dd4a-23e6-4617-ae86-d0cc64bec9d8\") " pod="openstack/openstackclient" Oct 07 14:04:03 crc kubenswrapper[4959]: I1007 14:04:03.309036 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4g8s\" (UniqueName: \"kubernetes.io/projected/9ba3dd4a-23e6-4617-ae86-d0cc64bec9d8-kube-api-access-s4g8s\") pod \"openstackclient\" (UID: \"9ba3dd4a-23e6-4617-ae86-d0cc64bec9d8\") " pod="openstack/openstackclient" Oct 07 14:04:03 crc kubenswrapper[4959]: I1007 14:04:03.388441 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 07 14:04:03 crc kubenswrapper[4959]: I1007 14:04:03.860445 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 07 14:04:03 crc kubenswrapper[4959]: W1007 14:04:03.861628 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ba3dd4a_23e6_4617_ae86_d0cc64bec9d8.slice/crio-f0d3bc4635847085b4beafa8e02a026ef06bf9aa352997a8e620ea95cccd7dcd WatchSource:0}: Error finding container f0d3bc4635847085b4beafa8e02a026ef06bf9aa352997a8e620ea95cccd7dcd: Status 404 returned error can't find the container with id f0d3bc4635847085b4beafa8e02a026ef06bf9aa352997a8e620ea95cccd7dcd Oct 07 14:04:04 crc kubenswrapper[4959]: I1007 14:04:04.146972 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"9ba3dd4a-23e6-4617-ae86-d0cc64bec9d8","Type":"ContainerStarted","Data":"f0d3bc4635847085b4beafa8e02a026ef06bf9aa352997a8e620ea95cccd7dcd"} Oct 07 14:04:04 crc kubenswrapper[4959]: I1007 14:04:04.149395 4959 generic.go:334] "Generic (PLEG): container finished" podID="a588fbdc-fd65-49f6-ab59-4901c7ca690a" containerID="76d664db4d06017fdcd093f711c6c672a5feacc3d610e94a71a01f0793720247" exitCode=0 Oct 07 14:04:04 crc kubenswrapper[4959]: I1007 14:04:04.149514 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-744bf8f46d-6twnn" event={"ID":"a588fbdc-fd65-49f6-ab59-4901c7ca690a","Type":"ContainerDied","Data":"76d664db4d06017fdcd093f711c6c672a5feacc3d610e94a71a01f0793720247"} Oct 07 14:04:04 crc kubenswrapper[4959]: I1007 14:04:04.688562 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 07 14:04:06 crc kubenswrapper[4959]: I1007 14:04:06.488941 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-744bf8f46d-6twnn" podUID="a588fbdc-fd65-49f6-ab59-4901c7ca690a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Oct 07 14:04:09 crc kubenswrapper[4959]: I1007 14:04:09.071748 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-d8944fbd8-5mpd2" Oct 07 14:04:09 crc kubenswrapper[4959]: I1007 14:04:09.379303 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-jmxbs"] Oct 07 14:04:09 crc kubenswrapper[4959]: I1007 14:04:09.380378 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-jmxbs" Oct 07 14:04:09 crc kubenswrapper[4959]: I1007 14:04:09.390728 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-jmxbs"] Oct 07 14:04:09 crc kubenswrapper[4959]: I1007 14:04:09.401327 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bdnz\" (UniqueName: \"kubernetes.io/projected/7ab681da-c60c-415b-985c-ca36c4cbead0-kube-api-access-5bdnz\") pod \"nova-api-db-create-jmxbs\" (UID: \"7ab681da-c60c-415b-985c-ca36c4cbead0\") " pod="openstack/nova-api-db-create-jmxbs" Oct 07 14:04:09 crc kubenswrapper[4959]: I1007 14:04:09.478772 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-77f42"] Oct 07 14:04:09 crc kubenswrapper[4959]: I1007 14:04:09.479798 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-77f42" Oct 07 14:04:09 crc kubenswrapper[4959]: I1007 14:04:09.489692 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-77f42"] Oct 07 14:04:09 crc kubenswrapper[4959]: I1007 14:04:09.503482 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bdnz\" (UniqueName: \"kubernetes.io/projected/7ab681da-c60c-415b-985c-ca36c4cbead0-kube-api-access-5bdnz\") pod \"nova-api-db-create-jmxbs\" (UID: \"7ab681da-c60c-415b-985c-ca36c4cbead0\") " pod="openstack/nova-api-db-create-jmxbs" Oct 07 14:04:09 crc kubenswrapper[4959]: I1007 14:04:09.504330 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zcp2\" (UniqueName: \"kubernetes.io/projected/861069fb-97b5-4ae8-a7da-0f0c51ff50ac-kube-api-access-9zcp2\") pod \"nova-cell0-db-create-77f42\" (UID: \"861069fb-97b5-4ae8-a7da-0f0c51ff50ac\") " pod="openstack/nova-cell0-db-create-77f42" Oct 07 14:04:09 crc kubenswrapper[4959]: I1007 14:04:09.529183 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bdnz\" (UniqueName: \"kubernetes.io/projected/7ab681da-c60c-415b-985c-ca36c4cbead0-kube-api-access-5bdnz\") pod \"nova-api-db-create-jmxbs\" (UID: \"7ab681da-c60c-415b-985c-ca36c4cbead0\") " pod="openstack/nova-api-db-create-jmxbs" Oct 07 14:04:09 crc kubenswrapper[4959]: I1007 14:04:09.577272 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-g494x"] Oct 07 14:04:09 crc kubenswrapper[4959]: I1007 14:04:09.578660 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-g494x" Oct 07 14:04:09 crc kubenswrapper[4959]: I1007 14:04:09.583328 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-g494x"] Oct 07 14:04:09 crc kubenswrapper[4959]: I1007 14:04:09.606333 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zcp2\" (UniqueName: \"kubernetes.io/projected/861069fb-97b5-4ae8-a7da-0f0c51ff50ac-kube-api-access-9zcp2\") pod \"nova-cell0-db-create-77f42\" (UID: \"861069fb-97b5-4ae8-a7da-0f0c51ff50ac\") " pod="openstack/nova-cell0-db-create-77f42" Oct 07 14:04:09 crc kubenswrapper[4959]: I1007 14:04:09.606481 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnhw2\" (UniqueName: \"kubernetes.io/projected/bb1f3a07-1ea0-4d4d-aba7-bc1912e3a47c-kube-api-access-xnhw2\") pod \"nova-cell1-db-create-g494x\" (UID: \"bb1f3a07-1ea0-4d4d-aba7-bc1912e3a47c\") " pod="openstack/nova-cell1-db-create-g494x" Oct 07 14:04:09 crc kubenswrapper[4959]: I1007 14:04:09.624650 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zcp2\" (UniqueName: \"kubernetes.io/projected/861069fb-97b5-4ae8-a7da-0f0c51ff50ac-kube-api-access-9zcp2\") pod \"nova-cell0-db-create-77f42\" (UID: \"861069fb-97b5-4ae8-a7da-0f0c51ff50ac\") " pod="openstack/nova-cell0-db-create-77f42" Oct 07 14:04:09 crc kubenswrapper[4959]: I1007 14:04:09.673421 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:04:09 crc kubenswrapper[4959]: I1007 14:04:09.673701 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bdf1b525-37e6-46c8-b972-3750fa675cd7" containerName="ceilometer-central-agent" containerID="cri-o://4eaf793c7493afd3e19b0fce8360a7d6fb26b4001a687df59ddd8cf1af2c649a" gracePeriod=30 Oct 07 14:04:09 crc kubenswrapper[4959]: I1007 14:04:09.673984 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bdf1b525-37e6-46c8-b972-3750fa675cd7" containerName="proxy-httpd" containerID="cri-o://d22923d6cbb6d0255c3d2dc8491f0609142a3434fac743bd79f54081c868bb66" gracePeriod=30 Oct 07 14:04:09 crc kubenswrapper[4959]: I1007 14:04:09.674007 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bdf1b525-37e6-46c8-b972-3750fa675cd7" containerName="sg-core" containerID="cri-o://da85e0da09a1be48d799f0a5d54ac72d60b28feed7cfb637414b5d2719b1e0c4" gracePeriod=30 Oct 07 14:04:09 crc kubenswrapper[4959]: I1007 14:04:09.674019 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bdf1b525-37e6-46c8-b972-3750fa675cd7" containerName="ceilometer-notification-agent" containerID="cri-o://8efefa3d8491a9f20e21845146780f1e5a3dd93f0d529d4c9253d4f128136913" gracePeriod=30 Oct 07 14:04:09 crc kubenswrapper[4959]: I1007 14:04:09.680535 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 07 14:04:09 crc kubenswrapper[4959]: I1007 14:04:09.700409 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-jmxbs" Oct 07 14:04:09 crc kubenswrapper[4959]: I1007 14:04:09.710628 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnhw2\" (UniqueName: \"kubernetes.io/projected/bb1f3a07-1ea0-4d4d-aba7-bc1912e3a47c-kube-api-access-xnhw2\") pod \"nova-cell1-db-create-g494x\" (UID: \"bb1f3a07-1ea0-4d4d-aba7-bc1912e3a47c\") " pod="openstack/nova-cell1-db-create-g494x" Oct 07 14:04:09 crc kubenswrapper[4959]: I1007 14:04:09.731746 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnhw2\" (UniqueName: \"kubernetes.io/projected/bb1f3a07-1ea0-4d4d-aba7-bc1912e3a47c-kube-api-access-xnhw2\") pod \"nova-cell1-db-create-g494x\" (UID: \"bb1f3a07-1ea0-4d4d-aba7-bc1912e3a47c\") " pod="openstack/nova-cell1-db-create-g494x" Oct 07 14:04:09 crc kubenswrapper[4959]: I1007 14:04:09.799705 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-77f42" Oct 07 14:04:09 crc kubenswrapper[4959]: I1007 14:04:09.913336 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-g494x" Oct 07 14:04:10 crc kubenswrapper[4959]: I1007 14:04:10.206800 4959 generic.go:334] "Generic (PLEG): container finished" podID="bdf1b525-37e6-46c8-b972-3750fa675cd7" containerID="d22923d6cbb6d0255c3d2dc8491f0609142a3434fac743bd79f54081c868bb66" exitCode=0 Oct 07 14:04:10 crc kubenswrapper[4959]: I1007 14:04:10.206831 4959 generic.go:334] "Generic (PLEG): container finished" podID="bdf1b525-37e6-46c8-b972-3750fa675cd7" containerID="da85e0da09a1be48d799f0a5d54ac72d60b28feed7cfb637414b5d2719b1e0c4" exitCode=2 Oct 07 14:04:10 crc kubenswrapper[4959]: I1007 14:04:10.206838 4959 generic.go:334] "Generic (PLEG): container finished" podID="bdf1b525-37e6-46c8-b972-3750fa675cd7" containerID="4eaf793c7493afd3e19b0fce8360a7d6fb26b4001a687df59ddd8cf1af2c649a" exitCode=0 Oct 07 14:04:10 crc kubenswrapper[4959]: I1007 14:04:10.206856 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdf1b525-37e6-46c8-b972-3750fa675cd7","Type":"ContainerDied","Data":"d22923d6cbb6d0255c3d2dc8491f0609142a3434fac743bd79f54081c868bb66"} Oct 07 14:04:10 crc kubenswrapper[4959]: I1007 14:04:10.206879 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdf1b525-37e6-46c8-b972-3750fa675cd7","Type":"ContainerDied","Data":"da85e0da09a1be48d799f0a5d54ac72d60b28feed7cfb637414b5d2719b1e0c4"} Oct 07 14:04:10 crc kubenswrapper[4959]: I1007 14:04:10.206889 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdf1b525-37e6-46c8-b972-3750fa675cd7","Type":"ContainerDied","Data":"4eaf793c7493afd3e19b0fce8360a7d6fb26b4001a687df59ddd8cf1af2c649a"} Oct 07 14:04:12 crc kubenswrapper[4959]: I1007 14:04:12.213279 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="bdf1b525-37e6-46c8-b972-3750fa675cd7" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.158:3000/\": dial tcp 10.217.0.158:3000: connect: connection refused" Oct 07 14:04:12 crc kubenswrapper[4959]: I1007 14:04:12.222420 4959 generic.go:334] "Generic (PLEG): container finished" podID="bdf1b525-37e6-46c8-b972-3750fa675cd7" containerID="8efefa3d8491a9f20e21845146780f1e5a3dd93f0d529d4c9253d4f128136913" exitCode=0 Oct 07 14:04:12 crc kubenswrapper[4959]: I1007 14:04:12.222456 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdf1b525-37e6-46c8-b972-3750fa675cd7","Type":"ContainerDied","Data":"8efefa3d8491a9f20e21845146780f1e5a3dd93f0d529d4c9253d4f128136913"} Oct 07 14:04:13 crc kubenswrapper[4959]: I1007 14:04:13.378128 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:04:13 crc kubenswrapper[4959]: I1007 14:04:13.480719 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdf1b525-37e6-46c8-b972-3750fa675cd7-config-data\") pod \"bdf1b525-37e6-46c8-b972-3750fa675cd7\" (UID: \"bdf1b525-37e6-46c8-b972-3750fa675cd7\") " Oct 07 14:04:13 crc kubenswrapper[4959]: I1007 14:04:13.480789 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2m9d\" (UniqueName: \"kubernetes.io/projected/bdf1b525-37e6-46c8-b972-3750fa675cd7-kube-api-access-c2m9d\") pod \"bdf1b525-37e6-46c8-b972-3750fa675cd7\" (UID: \"bdf1b525-37e6-46c8-b972-3750fa675cd7\") " Oct 07 14:04:13 crc kubenswrapper[4959]: I1007 14:04:13.480859 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdf1b525-37e6-46c8-b972-3750fa675cd7-scripts\") pod \"bdf1b525-37e6-46c8-b972-3750fa675cd7\" (UID: \"bdf1b525-37e6-46c8-b972-3750fa675cd7\") " Oct 07 14:04:13 crc kubenswrapper[4959]: I1007 14:04:13.480954 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdf1b525-37e6-46c8-b972-3750fa675cd7-combined-ca-bundle\") pod \"bdf1b525-37e6-46c8-b972-3750fa675cd7\" (UID: \"bdf1b525-37e6-46c8-b972-3750fa675cd7\") " Oct 07 14:04:13 crc kubenswrapper[4959]: I1007 14:04:13.481000 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdf1b525-37e6-46c8-b972-3750fa675cd7-run-httpd\") pod \"bdf1b525-37e6-46c8-b972-3750fa675cd7\" (UID: \"bdf1b525-37e6-46c8-b972-3750fa675cd7\") " Oct 07 14:04:13 crc kubenswrapper[4959]: I1007 14:04:13.481019 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdf1b525-37e6-46c8-b972-3750fa675cd7-log-httpd\") pod \"bdf1b525-37e6-46c8-b972-3750fa675cd7\" (UID: \"bdf1b525-37e6-46c8-b972-3750fa675cd7\") " Oct 07 14:04:13 crc kubenswrapper[4959]: I1007 14:04:13.481072 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bdf1b525-37e6-46c8-b972-3750fa675cd7-sg-core-conf-yaml\") pod \"bdf1b525-37e6-46c8-b972-3750fa675cd7\" (UID: \"bdf1b525-37e6-46c8-b972-3750fa675cd7\") " Oct 07 14:04:13 crc kubenswrapper[4959]: I1007 14:04:13.481790 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdf1b525-37e6-46c8-b972-3750fa675cd7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bdf1b525-37e6-46c8-b972-3750fa675cd7" (UID: "bdf1b525-37e6-46c8-b972-3750fa675cd7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:04:13 crc kubenswrapper[4959]: I1007 14:04:13.482164 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdf1b525-37e6-46c8-b972-3750fa675cd7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bdf1b525-37e6-46c8-b972-3750fa675cd7" (UID: "bdf1b525-37e6-46c8-b972-3750fa675cd7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:04:13 crc kubenswrapper[4959]: I1007 14:04:13.482535 4959 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdf1b525-37e6-46c8-b972-3750fa675cd7-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:13 crc kubenswrapper[4959]: I1007 14:04:13.482553 4959 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdf1b525-37e6-46c8-b972-3750fa675cd7-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:13 crc kubenswrapper[4959]: I1007 14:04:13.488223 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdf1b525-37e6-46c8-b972-3750fa675cd7-scripts" (OuterVolumeSpecName: "scripts") pod "bdf1b525-37e6-46c8-b972-3750fa675cd7" (UID: "bdf1b525-37e6-46c8-b972-3750fa675cd7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:04:13 crc kubenswrapper[4959]: I1007 14:04:13.491443 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdf1b525-37e6-46c8-b972-3750fa675cd7-kube-api-access-c2m9d" (OuterVolumeSpecName: "kube-api-access-c2m9d") pod "bdf1b525-37e6-46c8-b972-3750fa675cd7" (UID: "bdf1b525-37e6-46c8-b972-3750fa675cd7"). InnerVolumeSpecName "kube-api-access-c2m9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:04:13 crc kubenswrapper[4959]: I1007 14:04:13.526254 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdf1b525-37e6-46c8-b972-3750fa675cd7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bdf1b525-37e6-46c8-b972-3750fa675cd7" (UID: "bdf1b525-37e6-46c8-b972-3750fa675cd7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:04:13 crc kubenswrapper[4959]: I1007 14:04:13.564026 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdf1b525-37e6-46c8-b972-3750fa675cd7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bdf1b525-37e6-46c8-b972-3750fa675cd7" (UID: "bdf1b525-37e6-46c8-b972-3750fa675cd7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:04:13 crc kubenswrapper[4959]: I1007 14:04:13.586776 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2m9d\" (UniqueName: \"kubernetes.io/projected/bdf1b525-37e6-46c8-b972-3750fa675cd7-kube-api-access-c2m9d\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:13 crc kubenswrapper[4959]: I1007 14:04:13.586808 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdf1b525-37e6-46c8-b972-3750fa675cd7-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:13 crc kubenswrapper[4959]: I1007 14:04:13.586818 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdf1b525-37e6-46c8-b972-3750fa675cd7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:13 crc kubenswrapper[4959]: I1007 14:04:13.586828 4959 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bdf1b525-37e6-46c8-b972-3750fa675cd7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:13 crc kubenswrapper[4959]: I1007 14:04:13.601077 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-g494x"] Oct 07 14:04:13 crc kubenswrapper[4959]: I1007 14:04:13.609662 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-77f42"] Oct 07 14:04:13 crc kubenswrapper[4959]: I1007 14:04:13.625745 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdf1b525-37e6-46c8-b972-3750fa675cd7-config-data" (OuterVolumeSpecName: "config-data") pod "bdf1b525-37e6-46c8-b972-3750fa675cd7" (UID: "bdf1b525-37e6-46c8-b972-3750fa675cd7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:04:13 crc kubenswrapper[4959]: I1007 14:04:13.626329 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-jmxbs"] Oct 07 14:04:13 crc kubenswrapper[4959]: I1007 14:04:13.688200 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdf1b525-37e6-46c8-b972-3750fa675cd7-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.252898 4959 generic.go:334] "Generic (PLEG): container finished" podID="bb1f3a07-1ea0-4d4d-aba7-bc1912e3a47c" containerID="23adf200733196bc33a98c2ba5078225882e47f27d4fad79ecbc95256ca18655" exitCode=0 Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.252976 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-g494x" event={"ID":"bb1f3a07-1ea0-4d4d-aba7-bc1912e3a47c","Type":"ContainerDied","Data":"23adf200733196bc33a98c2ba5078225882e47f27d4fad79ecbc95256ca18655"} Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.253240 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-g494x" event={"ID":"bb1f3a07-1ea0-4d4d-aba7-bc1912e3a47c","Type":"ContainerStarted","Data":"c0d99d10cb89d8b93e4325a392a83c9c94e58487eda07280ef6770e2531f7212"} Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.257264 4959 generic.go:334] "Generic (PLEG): container finished" podID="7ab681da-c60c-415b-985c-ca36c4cbead0" containerID="8516f5e1fc246fe52543f3db0e6133ecf48e28d79a57e9be359176aea738d69f" exitCode=0 Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.257380 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-jmxbs" event={"ID":"7ab681da-c60c-415b-985c-ca36c4cbead0","Type":"ContainerDied","Data":"8516f5e1fc246fe52543f3db0e6133ecf48e28d79a57e9be359176aea738d69f"} Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.257418 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-jmxbs" event={"ID":"7ab681da-c60c-415b-985c-ca36c4cbead0","Type":"ContainerStarted","Data":"b23b35ed631dc26adbed1a9f274c03f37bef326d06784f5598ce0a7f36e35dbe"} Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.286399 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.286641 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdf1b525-37e6-46c8-b972-3750fa675cd7","Type":"ContainerDied","Data":"f8a8c98d53a271028e1c810511fa6e9b724a84d3a84af6a4ea99228858c6409e"} Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.286681 4959 scope.go:117] "RemoveContainer" containerID="d22923d6cbb6d0255c3d2dc8491f0609142a3434fac743bd79f54081c868bb66" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.289021 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"9ba3dd4a-23e6-4617-ae86-d0cc64bec9d8","Type":"ContainerStarted","Data":"e5c3da1b8100b5a2725927f31174e477d40000b5eff13a9fb7dcb1331b31de67"} Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.291831 4959 generic.go:334] "Generic (PLEG): container finished" podID="861069fb-97b5-4ae8-a7da-0f0c51ff50ac" containerID="a838210ef105ae5f335860c3556e14b59e63b90c800bb0fc31efab2ec05db490" exitCode=0 Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.291863 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-77f42" event={"ID":"861069fb-97b5-4ae8-a7da-0f0c51ff50ac","Type":"ContainerDied","Data":"a838210ef105ae5f335860c3556e14b59e63b90c800bb0fc31efab2ec05db490"} Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.291878 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-77f42" event={"ID":"861069fb-97b5-4ae8-a7da-0f0c51ff50ac","Type":"ContainerStarted","Data":"1abde5574dc9a90d78dc4d686231447aa08337040f8cbba3f25d7527ae3ca8bd"} Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.311228 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.182841247 podStartE2EDuration="11.311208558s" podCreationTimestamp="2025-10-07 14:04:03 +0000 UTC" firstStartedPulling="2025-10-07 14:04:03.86382939 +0000 UTC m=+1125.947233705" lastFinishedPulling="2025-10-07 14:04:12.992196681 +0000 UTC m=+1135.075601016" observedRunningTime="2025-10-07 14:04:14.301397395 +0000 UTC m=+1136.384801720" watchObservedRunningTime="2025-10-07 14:04:14.311208558 +0000 UTC m=+1136.394612883" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.313127 4959 scope.go:117] "RemoveContainer" containerID="da85e0da09a1be48d799f0a5d54ac72d60b28feed7cfb637414b5d2719b1e0c4" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.338684 4959 scope.go:117] "RemoveContainer" containerID="8efefa3d8491a9f20e21845146780f1e5a3dd93f0d529d4c9253d4f128136913" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.346789 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.372279 4959 scope.go:117] "RemoveContainer" containerID="4eaf793c7493afd3e19b0fce8360a7d6fb26b4001a687df59ddd8cf1af2c649a" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.373128 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.394753 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:04:14 crc kubenswrapper[4959]: E1007 14:04:14.395187 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdf1b525-37e6-46c8-b972-3750fa675cd7" containerName="ceilometer-central-agent" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.395203 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdf1b525-37e6-46c8-b972-3750fa675cd7" containerName="ceilometer-central-agent" Oct 07 14:04:14 crc kubenswrapper[4959]: E1007 14:04:14.395222 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdf1b525-37e6-46c8-b972-3750fa675cd7" containerName="sg-core" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.395228 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdf1b525-37e6-46c8-b972-3750fa675cd7" containerName="sg-core" Oct 07 14:04:14 crc kubenswrapper[4959]: E1007 14:04:14.395256 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdf1b525-37e6-46c8-b972-3750fa675cd7" containerName="ceilometer-notification-agent" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.395262 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdf1b525-37e6-46c8-b972-3750fa675cd7" containerName="ceilometer-notification-agent" Oct 07 14:04:14 crc kubenswrapper[4959]: E1007 14:04:14.395278 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdf1b525-37e6-46c8-b972-3750fa675cd7" containerName="proxy-httpd" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.395283 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdf1b525-37e6-46c8-b972-3750fa675cd7" containerName="proxy-httpd" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.395503 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdf1b525-37e6-46c8-b972-3750fa675cd7" containerName="proxy-httpd" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.395521 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdf1b525-37e6-46c8-b972-3750fa675cd7" containerName="sg-core" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.395533 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdf1b525-37e6-46c8-b972-3750fa675cd7" containerName="ceilometer-central-agent" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.395542 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdf1b525-37e6-46c8-b972-3750fa675cd7" containerName="ceilometer-notification-agent" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.397269 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.401165 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.401270 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.401330 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.501119 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prvsh\" (UniqueName: \"kubernetes.io/projected/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-kube-api-access-prvsh\") pod \"ceilometer-0\" (UID: \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\") " pod="openstack/ceilometer-0" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.501893 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\") " pod="openstack/ceilometer-0" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.501970 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-run-httpd\") pod \"ceilometer-0\" (UID: \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\") " pod="openstack/ceilometer-0" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.502032 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-log-httpd\") pod \"ceilometer-0\" (UID: \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\") " pod="openstack/ceilometer-0" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.502073 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\") " pod="openstack/ceilometer-0" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.502307 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-config-data\") pod \"ceilometer-0\" (UID: \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\") " pod="openstack/ceilometer-0" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.502365 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-scripts\") pod \"ceilometer-0\" (UID: \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\") " pod="openstack/ceilometer-0" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.603477 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\") " pod="openstack/ceilometer-0" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.603564 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-run-httpd\") pod \"ceilometer-0\" (UID: \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\") " pod="openstack/ceilometer-0" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.603606 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-log-httpd\") pod \"ceilometer-0\" (UID: \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\") " pod="openstack/ceilometer-0" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.603639 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\") " pod="openstack/ceilometer-0" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.603696 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-config-data\") pod \"ceilometer-0\" (UID: \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\") " pod="openstack/ceilometer-0" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.603717 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-scripts\") pod \"ceilometer-0\" (UID: \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\") " pod="openstack/ceilometer-0" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.603777 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prvsh\" (UniqueName: \"kubernetes.io/projected/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-kube-api-access-prvsh\") pod \"ceilometer-0\" (UID: \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\") " pod="openstack/ceilometer-0" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.605177 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-run-httpd\") pod \"ceilometer-0\" (UID: \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\") " pod="openstack/ceilometer-0" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.605290 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-log-httpd\") pod \"ceilometer-0\" (UID: \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\") " pod="openstack/ceilometer-0" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.609366 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-config-data\") pod \"ceilometer-0\" (UID: \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\") " pod="openstack/ceilometer-0" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.611006 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\") " pod="openstack/ceilometer-0" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.619313 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prvsh\" (UniqueName: \"kubernetes.io/projected/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-kube-api-access-prvsh\") pod \"ceilometer-0\" (UID: \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\") " pod="openstack/ceilometer-0" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.626744 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-scripts\") pod \"ceilometer-0\" (UID: \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\") " pod="openstack/ceilometer-0" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.627031 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\") " pod="openstack/ceilometer-0" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.664815 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdf1b525-37e6-46c8-b972-3750fa675cd7" path="/var/lib/kubelet/pods/bdf1b525-37e6-46c8-b972-3750fa675cd7/volumes" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.718450 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:04:14 crc kubenswrapper[4959]: I1007 14:04:14.932994 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-67d6886b4f-lz5gr" Oct 07 14:04:15 crc kubenswrapper[4959]: I1007 14:04:15.004179 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-d8944fbd8-5mpd2"] Oct 07 14:04:15 crc kubenswrapper[4959]: I1007 14:04:15.006640 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-d8944fbd8-5mpd2" podUID="5142bfe0-aa71-4bf7-ad55-7264d0e6a94a" containerName="neutron-api" containerID="cri-o://2da84f09dcd2b60b3aba0ee6f3cf7ede92179d1cd800c821fc83723c68223df4" gracePeriod=30 Oct 07 14:04:15 crc kubenswrapper[4959]: I1007 14:04:15.006965 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-d8944fbd8-5mpd2" podUID="5142bfe0-aa71-4bf7-ad55-7264d0e6a94a" containerName="neutron-httpd" containerID="cri-o://0db0b06172375d883555d2f1bbb13a63749e39e63b474637ee52f4f21c685469" gracePeriod=30 Oct 07 14:04:15 crc kubenswrapper[4959]: W1007 14:04:15.179433 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b99319e_4fc6_4cfc_939c_14dd9a04ea21.slice/crio-22f8c634268c3e3303a3b67d0c33691e87b927ee0823fc174e7eb000dbe06fea WatchSource:0}: Error finding container 22f8c634268c3e3303a3b67d0c33691e87b927ee0823fc174e7eb000dbe06fea: Status 404 returned error can't find the container with id 22f8c634268c3e3303a3b67d0c33691e87b927ee0823fc174e7eb000dbe06fea Oct 07 14:04:15 crc kubenswrapper[4959]: I1007 14:04:15.181572 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:04:15 crc kubenswrapper[4959]: I1007 14:04:15.303065 4959 generic.go:334] "Generic (PLEG): container finished" podID="5142bfe0-aa71-4bf7-ad55-7264d0e6a94a" containerID="0db0b06172375d883555d2f1bbb13a63749e39e63b474637ee52f4f21c685469" exitCode=0 Oct 07 14:04:15 crc kubenswrapper[4959]: I1007 14:04:15.303119 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d8944fbd8-5mpd2" event={"ID":"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a","Type":"ContainerDied","Data":"0db0b06172375d883555d2f1bbb13a63749e39e63b474637ee52f4f21c685469"} Oct 07 14:04:15 crc kubenswrapper[4959]: I1007 14:04:15.306390 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b99319e-4fc6-4cfc-939c-14dd9a04ea21","Type":"ContainerStarted","Data":"22f8c634268c3e3303a3b67d0c33691e87b927ee0823fc174e7eb000dbe06fea"} Oct 07 14:04:15 crc kubenswrapper[4959]: I1007 14:04:15.675338 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-77f42" Oct 07 14:04:15 crc kubenswrapper[4959]: I1007 14:04:15.761404 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-jmxbs" Oct 07 14:04:15 crc kubenswrapper[4959]: I1007 14:04:15.770620 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-g494x" Oct 07 14:04:15 crc kubenswrapper[4959]: I1007 14:04:15.827172 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zcp2\" (UniqueName: \"kubernetes.io/projected/861069fb-97b5-4ae8-a7da-0f0c51ff50ac-kube-api-access-9zcp2\") pod \"861069fb-97b5-4ae8-a7da-0f0c51ff50ac\" (UID: \"861069fb-97b5-4ae8-a7da-0f0c51ff50ac\") " Oct 07 14:04:15 crc kubenswrapper[4959]: I1007 14:04:15.831634 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/861069fb-97b5-4ae8-a7da-0f0c51ff50ac-kube-api-access-9zcp2" (OuterVolumeSpecName: "kube-api-access-9zcp2") pod "861069fb-97b5-4ae8-a7da-0f0c51ff50ac" (UID: "861069fb-97b5-4ae8-a7da-0f0c51ff50ac"). InnerVolumeSpecName "kube-api-access-9zcp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:04:15 crc kubenswrapper[4959]: I1007 14:04:15.928549 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bdnz\" (UniqueName: \"kubernetes.io/projected/7ab681da-c60c-415b-985c-ca36c4cbead0-kube-api-access-5bdnz\") pod \"7ab681da-c60c-415b-985c-ca36c4cbead0\" (UID: \"7ab681da-c60c-415b-985c-ca36c4cbead0\") " Oct 07 14:04:15 crc kubenswrapper[4959]: I1007 14:04:15.928659 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnhw2\" (UniqueName: \"kubernetes.io/projected/bb1f3a07-1ea0-4d4d-aba7-bc1912e3a47c-kube-api-access-xnhw2\") pod \"bb1f3a07-1ea0-4d4d-aba7-bc1912e3a47c\" (UID: \"bb1f3a07-1ea0-4d4d-aba7-bc1912e3a47c\") " Oct 07 14:04:15 crc kubenswrapper[4959]: I1007 14:04:15.929294 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zcp2\" (UniqueName: \"kubernetes.io/projected/861069fb-97b5-4ae8-a7da-0f0c51ff50ac-kube-api-access-9zcp2\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:15 crc kubenswrapper[4959]: I1007 14:04:15.932397 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb1f3a07-1ea0-4d4d-aba7-bc1912e3a47c-kube-api-access-xnhw2" (OuterVolumeSpecName: "kube-api-access-xnhw2") pod "bb1f3a07-1ea0-4d4d-aba7-bc1912e3a47c" (UID: "bb1f3a07-1ea0-4d4d-aba7-bc1912e3a47c"). InnerVolumeSpecName "kube-api-access-xnhw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:04:15 crc kubenswrapper[4959]: I1007 14:04:15.933399 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ab681da-c60c-415b-985c-ca36c4cbead0-kube-api-access-5bdnz" (OuterVolumeSpecName: "kube-api-access-5bdnz") pod "7ab681da-c60c-415b-985c-ca36c4cbead0" (UID: "7ab681da-c60c-415b-985c-ca36c4cbead0"). InnerVolumeSpecName "kube-api-access-5bdnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:04:16 crc kubenswrapper[4959]: I1007 14:04:16.030379 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bdnz\" (UniqueName: \"kubernetes.io/projected/7ab681da-c60c-415b-985c-ca36c4cbead0-kube-api-access-5bdnz\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:16 crc kubenswrapper[4959]: I1007 14:04:16.030660 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnhw2\" (UniqueName: \"kubernetes.io/projected/bb1f3a07-1ea0-4d4d-aba7-bc1912e3a47c-kube-api-access-xnhw2\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:16 crc kubenswrapper[4959]: I1007 14:04:16.321182 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b99319e-4fc6-4cfc-939c-14dd9a04ea21","Type":"ContainerStarted","Data":"eb81d79bfd2cc50651cd8c8486b6ec8d06cc2e33553e74fda7ad7a680ae04dbb"} Oct 07 14:04:16 crc kubenswrapper[4959]: I1007 14:04:16.324465 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-g494x" Oct 07 14:04:16 crc kubenswrapper[4959]: I1007 14:04:16.324743 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-g494x" event={"ID":"bb1f3a07-1ea0-4d4d-aba7-bc1912e3a47c","Type":"ContainerDied","Data":"c0d99d10cb89d8b93e4325a392a83c9c94e58487eda07280ef6770e2531f7212"} Oct 07 14:04:16 crc kubenswrapper[4959]: I1007 14:04:16.324787 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0d99d10cb89d8b93e4325a392a83c9c94e58487eda07280ef6770e2531f7212" Oct 07 14:04:16 crc kubenswrapper[4959]: I1007 14:04:16.328614 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-jmxbs" event={"ID":"7ab681da-c60c-415b-985c-ca36c4cbead0","Type":"ContainerDied","Data":"b23b35ed631dc26adbed1a9f274c03f37bef326d06784f5598ce0a7f36e35dbe"} Oct 07 14:04:16 crc kubenswrapper[4959]: I1007 14:04:16.328642 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b23b35ed631dc26adbed1a9f274c03f37bef326d06784f5598ce0a7f36e35dbe" Oct 07 14:04:16 crc kubenswrapper[4959]: I1007 14:04:16.328697 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-jmxbs" Oct 07 14:04:16 crc kubenswrapper[4959]: I1007 14:04:16.336028 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-77f42" event={"ID":"861069fb-97b5-4ae8-a7da-0f0c51ff50ac","Type":"ContainerDied","Data":"1abde5574dc9a90d78dc4d686231447aa08337040f8cbba3f25d7527ae3ca8bd"} Oct 07 14:04:16 crc kubenswrapper[4959]: I1007 14:04:16.336075 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1abde5574dc9a90d78dc4d686231447aa08337040f8cbba3f25d7527ae3ca8bd" Oct 07 14:04:16 crc kubenswrapper[4959]: I1007 14:04:16.336154 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-77f42" Oct 07 14:04:16 crc kubenswrapper[4959]: I1007 14:04:16.489750 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-744bf8f46d-6twnn" podUID="a588fbdc-fd65-49f6-ab59-4901c7ca690a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Oct 07 14:04:16 crc kubenswrapper[4959]: I1007 14:04:16.522009 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:04:18 crc kubenswrapper[4959]: I1007 14:04:18.367773 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b99319e-4fc6-4cfc-939c-14dd9a04ea21","Type":"ContainerStarted","Data":"8d9e92ff000993b1e005b5c4259b07c490aa0a2badf1c746cb1d741de3aca253"} Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.261750 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d8944fbd8-5mpd2" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.381489 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b99319e-4fc6-4cfc-939c-14dd9a04ea21","Type":"ContainerStarted","Data":"1fd6a48cf687a2cbaea1ddac89b0a3b3b2baf2eac2d793e8a0d86a733826359c"} Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.383930 4959 generic.go:334] "Generic (PLEG): container finished" podID="5142bfe0-aa71-4bf7-ad55-7264d0e6a94a" containerID="2da84f09dcd2b60b3aba0ee6f3cf7ede92179d1cd800c821fc83723c68223df4" exitCode=0 Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.383968 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d8944fbd8-5mpd2" event={"ID":"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a","Type":"ContainerDied","Data":"2da84f09dcd2b60b3aba0ee6f3cf7ede92179d1cd800c821fc83723c68223df4"} Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.383989 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d8944fbd8-5mpd2" event={"ID":"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a","Type":"ContainerDied","Data":"2568bafb261674b5e4117efa8c6dcc6a1a271e9b1fca93456822da5dfd6f4693"} Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.384009 4959 scope.go:117] "RemoveContainer" containerID="0db0b06172375d883555d2f1bbb13a63749e39e63b474637ee52f4f21c685469" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.384194 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d8944fbd8-5mpd2" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.409282 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-combined-ca-bundle\") pod \"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a\" (UID: \"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a\") " Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.409374 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-ovndb-tls-certs\") pod \"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a\" (UID: \"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a\") " Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.409434 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8kxg\" (UniqueName: \"kubernetes.io/projected/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-kube-api-access-s8kxg\") pod \"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a\" (UID: \"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a\") " Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.409477 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-httpd-config\") pod \"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a\" (UID: \"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a\") " Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.409503 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-config\") pod \"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a\" (UID: \"5142bfe0-aa71-4bf7-ad55-7264d0e6a94a\") " Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.414529 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "5142bfe0-aa71-4bf7-ad55-7264d0e6a94a" (UID: "5142bfe0-aa71-4bf7-ad55-7264d0e6a94a"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.415814 4959 scope.go:117] "RemoveContainer" containerID="2da84f09dcd2b60b3aba0ee6f3cf7ede92179d1cd800c821fc83723c68223df4" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.430186 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-kube-api-access-s8kxg" (OuterVolumeSpecName: "kube-api-access-s8kxg") pod "5142bfe0-aa71-4bf7-ad55-7264d0e6a94a" (UID: "5142bfe0-aa71-4bf7-ad55-7264d0e6a94a"). InnerVolumeSpecName "kube-api-access-s8kxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.462453 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5142bfe0-aa71-4bf7-ad55-7264d0e6a94a" (UID: "5142bfe0-aa71-4bf7-ad55-7264d0e6a94a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.486934 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-config" (OuterVolumeSpecName: "config") pod "5142bfe0-aa71-4bf7-ad55-7264d0e6a94a" (UID: "5142bfe0-aa71-4bf7-ad55-7264d0e6a94a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.512882 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.514323 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8kxg\" (UniqueName: \"kubernetes.io/projected/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-kube-api-access-s8kxg\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.517256 4959 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.517288 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-config\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.537902 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "5142bfe0-aa71-4bf7-ad55-7264d0e6a94a" (UID: "5142bfe0-aa71-4bf7-ad55-7264d0e6a94a"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.574428 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-3e47-account-create-h55ck"] Oct 07 14:04:19 crc kubenswrapper[4959]: E1007 14:04:19.574819 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb1f3a07-1ea0-4d4d-aba7-bc1912e3a47c" containerName="mariadb-database-create" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.574833 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb1f3a07-1ea0-4d4d-aba7-bc1912e3a47c" containerName="mariadb-database-create" Oct 07 14:04:19 crc kubenswrapper[4959]: E1007 14:04:19.574846 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ab681da-c60c-415b-985c-ca36c4cbead0" containerName="mariadb-database-create" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.574853 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ab681da-c60c-415b-985c-ca36c4cbead0" containerName="mariadb-database-create" Oct 07 14:04:19 crc kubenswrapper[4959]: E1007 14:04:19.574879 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5142bfe0-aa71-4bf7-ad55-7264d0e6a94a" containerName="neutron-api" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.574887 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5142bfe0-aa71-4bf7-ad55-7264d0e6a94a" containerName="neutron-api" Oct 07 14:04:19 crc kubenswrapper[4959]: E1007 14:04:19.574897 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5142bfe0-aa71-4bf7-ad55-7264d0e6a94a" containerName="neutron-httpd" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.574904 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5142bfe0-aa71-4bf7-ad55-7264d0e6a94a" containerName="neutron-httpd" Oct 07 14:04:19 crc kubenswrapper[4959]: E1007 14:04:19.574922 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="861069fb-97b5-4ae8-a7da-0f0c51ff50ac" containerName="mariadb-database-create" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.574929 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="861069fb-97b5-4ae8-a7da-0f0c51ff50ac" containerName="mariadb-database-create" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.575138 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="5142bfe0-aa71-4bf7-ad55-7264d0e6a94a" containerName="neutron-httpd" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.575157 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="5142bfe0-aa71-4bf7-ad55-7264d0e6a94a" containerName="neutron-api" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.575175 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="861069fb-97b5-4ae8-a7da-0f0c51ff50ac" containerName="mariadb-database-create" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.575186 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb1f3a07-1ea0-4d4d-aba7-bc1912e3a47c" containerName="mariadb-database-create" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.575196 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ab681da-c60c-415b-985c-ca36c4cbead0" containerName="mariadb-database-create" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.577028 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3e47-account-create-h55ck" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.585791 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.591426 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-3e47-account-create-h55ck"] Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.618378 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksvf5\" (UniqueName: \"kubernetes.io/projected/c9743ad9-caa9-4bc4-97ef-fa52706ca1ae-kube-api-access-ksvf5\") pod \"nova-api-3e47-account-create-h55ck\" (UID: \"c9743ad9-caa9-4bc4-97ef-fa52706ca1ae\") " pod="openstack/nova-api-3e47-account-create-h55ck" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.618507 4959 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.643277 4959 scope.go:117] "RemoveContainer" containerID="0db0b06172375d883555d2f1bbb13a63749e39e63b474637ee52f4f21c685469" Oct 07 14:04:19 crc kubenswrapper[4959]: E1007 14:04:19.647211 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0db0b06172375d883555d2f1bbb13a63749e39e63b474637ee52f4f21c685469\": container with ID starting with 0db0b06172375d883555d2f1bbb13a63749e39e63b474637ee52f4f21c685469 not found: ID does not exist" containerID="0db0b06172375d883555d2f1bbb13a63749e39e63b474637ee52f4f21c685469" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.647257 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0db0b06172375d883555d2f1bbb13a63749e39e63b474637ee52f4f21c685469"} err="failed to get container status \"0db0b06172375d883555d2f1bbb13a63749e39e63b474637ee52f4f21c685469\": rpc error: code = NotFound desc = could not find container \"0db0b06172375d883555d2f1bbb13a63749e39e63b474637ee52f4f21c685469\": container with ID starting with 0db0b06172375d883555d2f1bbb13a63749e39e63b474637ee52f4f21c685469 not found: ID does not exist" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.647288 4959 scope.go:117] "RemoveContainer" containerID="2da84f09dcd2b60b3aba0ee6f3cf7ede92179d1cd800c821fc83723c68223df4" Oct 07 14:04:19 crc kubenswrapper[4959]: E1007 14:04:19.647652 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2da84f09dcd2b60b3aba0ee6f3cf7ede92179d1cd800c821fc83723c68223df4\": container with ID starting with 2da84f09dcd2b60b3aba0ee6f3cf7ede92179d1cd800c821fc83723c68223df4 not found: ID does not exist" containerID="2da84f09dcd2b60b3aba0ee6f3cf7ede92179d1cd800c821fc83723c68223df4" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.647688 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2da84f09dcd2b60b3aba0ee6f3cf7ede92179d1cd800c821fc83723c68223df4"} err="failed to get container status \"2da84f09dcd2b60b3aba0ee6f3cf7ede92179d1cd800c821fc83723c68223df4\": rpc error: code = NotFound desc = could not find container \"2da84f09dcd2b60b3aba0ee6f3cf7ede92179d1cd800c821fc83723c68223df4\": container with ID starting with 2da84f09dcd2b60b3aba0ee6f3cf7ede92179d1cd800c821fc83723c68223df4 not found: ID does not exist" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.720215 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksvf5\" (UniqueName: \"kubernetes.io/projected/c9743ad9-caa9-4bc4-97ef-fa52706ca1ae-kube-api-access-ksvf5\") pod \"nova-api-3e47-account-create-h55ck\" (UID: \"c9743ad9-caa9-4bc4-97ef-fa52706ca1ae\") " pod="openstack/nova-api-3e47-account-create-h55ck" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.723453 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-d8944fbd8-5mpd2"] Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.729897 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-d8944fbd8-5mpd2"] Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.737375 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksvf5\" (UniqueName: \"kubernetes.io/projected/c9743ad9-caa9-4bc4-97ef-fa52706ca1ae-kube-api-access-ksvf5\") pod \"nova-api-3e47-account-create-h55ck\" (UID: \"c9743ad9-caa9-4bc4-97ef-fa52706ca1ae\") " pod="openstack/nova-api-3e47-account-create-h55ck" Oct 07 14:04:19 crc kubenswrapper[4959]: I1007 14:04:19.933797 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3e47-account-create-h55ck" Oct 07 14:04:20 crc kubenswrapper[4959]: I1007 14:04:20.378516 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-3e47-account-create-h55ck"] Oct 07 14:04:20 crc kubenswrapper[4959]: I1007 14:04:20.629904 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:04:20 crc kubenswrapper[4959]: I1007 14:04:20.629953 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:04:20 crc kubenswrapper[4959]: I1007 14:04:20.664124 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5142bfe0-aa71-4bf7-ad55-7264d0e6a94a" path="/var/lib/kubelet/pods/5142bfe0-aa71-4bf7-ad55-7264d0e6a94a/volumes" Oct 07 14:04:21 crc kubenswrapper[4959]: I1007 14:04:21.418924 4959 generic.go:334] "Generic (PLEG): container finished" podID="c9743ad9-caa9-4bc4-97ef-fa52706ca1ae" containerID="59fac583237ac7d0d6ba7ead48a081c6075c175f02ac7bbc456ae75dabbac082" exitCode=0 Oct 07 14:04:21 crc kubenswrapper[4959]: I1007 14:04:21.418966 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3e47-account-create-h55ck" event={"ID":"c9743ad9-caa9-4bc4-97ef-fa52706ca1ae","Type":"ContainerDied","Data":"59fac583237ac7d0d6ba7ead48a081c6075c175f02ac7bbc456ae75dabbac082"} Oct 07 14:04:21 crc kubenswrapper[4959]: I1007 14:04:21.419345 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3e47-account-create-h55ck" event={"ID":"c9743ad9-caa9-4bc4-97ef-fa52706ca1ae","Type":"ContainerStarted","Data":"445514157d1e163c9489d671e8c54cc1740606237a4cdf485113a5956ef0f864"} Oct 07 14:04:21 crc kubenswrapper[4959]: I1007 14:04:21.423282 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b99319e-4fc6-4cfc-939c-14dd9a04ea21","Type":"ContainerStarted","Data":"8b12ee5c290b974cdbb5d92113026cb92ba6b5dec8c5771a34c09a32d9197b14"} Oct 07 14:04:21 crc kubenswrapper[4959]: I1007 14:04:21.423428 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b99319e-4fc6-4cfc-939c-14dd9a04ea21" containerName="ceilometer-central-agent" containerID="cri-o://eb81d79bfd2cc50651cd8c8486b6ec8d06cc2e33553e74fda7ad7a680ae04dbb" gracePeriod=30 Oct 07 14:04:21 crc kubenswrapper[4959]: I1007 14:04:21.423516 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 07 14:04:21 crc kubenswrapper[4959]: I1007 14:04:21.423539 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b99319e-4fc6-4cfc-939c-14dd9a04ea21" containerName="sg-core" containerID="cri-o://1fd6a48cf687a2cbaea1ddac89b0a3b3b2baf2eac2d793e8a0d86a733826359c" gracePeriod=30 Oct 07 14:04:21 crc kubenswrapper[4959]: I1007 14:04:21.423581 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b99319e-4fc6-4cfc-939c-14dd9a04ea21" containerName="ceilometer-notification-agent" containerID="cri-o://8d9e92ff000993b1e005b5c4259b07c490aa0a2badf1c746cb1d741de3aca253" gracePeriod=30 Oct 07 14:04:21 crc kubenswrapper[4959]: I1007 14:04:21.423790 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b99319e-4fc6-4cfc-939c-14dd9a04ea21" containerName="proxy-httpd" containerID="cri-o://8b12ee5c290b974cdbb5d92113026cb92ba6b5dec8c5771a34c09a32d9197b14" gracePeriod=30 Oct 07 14:04:21 crc kubenswrapper[4959]: I1007 14:04:21.467861 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.347694598 podStartE2EDuration="7.467830593s" podCreationTimestamp="2025-10-07 14:04:14 +0000 UTC" firstStartedPulling="2025-10-07 14:04:15.184453578 +0000 UTC m=+1137.267857903" lastFinishedPulling="2025-10-07 14:04:20.304589583 +0000 UTC m=+1142.387993898" observedRunningTime="2025-10-07 14:04:21.458679077 +0000 UTC m=+1143.542083412" watchObservedRunningTime="2025-10-07 14:04:21.467830593 +0000 UTC m=+1143.551234918" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.125811 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.157186 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-log-httpd\") pod \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\" (UID: \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\") " Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.157235 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prvsh\" (UniqueName: \"kubernetes.io/projected/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-kube-api-access-prvsh\") pod \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\" (UID: \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\") " Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.157274 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-scripts\") pod \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\" (UID: \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\") " Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.157305 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-config-data\") pod \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\" (UID: \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\") " Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.157345 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-run-httpd\") pod \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\" (UID: \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\") " Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.157380 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-sg-core-conf-yaml\") pod \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\" (UID: \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\") " Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.157428 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-combined-ca-bundle\") pod \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\" (UID: \"5b99319e-4fc6-4cfc-939c-14dd9a04ea21\") " Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.161300 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5b99319e-4fc6-4cfc-939c-14dd9a04ea21" (UID: "5b99319e-4fc6-4cfc-939c-14dd9a04ea21"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.165214 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5b99319e-4fc6-4cfc-939c-14dd9a04ea21" (UID: "5b99319e-4fc6-4cfc-939c-14dd9a04ea21"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.168420 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-scripts" (OuterVolumeSpecName: "scripts") pod "5b99319e-4fc6-4cfc-939c-14dd9a04ea21" (UID: "5b99319e-4fc6-4cfc-939c-14dd9a04ea21"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.168456 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-kube-api-access-prvsh" (OuterVolumeSpecName: "kube-api-access-prvsh") pod "5b99319e-4fc6-4cfc-939c-14dd9a04ea21" (UID: "5b99319e-4fc6-4cfc-939c-14dd9a04ea21"). InnerVolumeSpecName "kube-api-access-prvsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.183935 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5b99319e-4fc6-4cfc-939c-14dd9a04ea21" (UID: "5b99319e-4fc6-4cfc-939c-14dd9a04ea21"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.236864 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b99319e-4fc6-4cfc-939c-14dd9a04ea21" (UID: "5b99319e-4fc6-4cfc-939c-14dd9a04ea21"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.241304 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-config-data" (OuterVolumeSpecName: "config-data") pod "5b99319e-4fc6-4cfc-939c-14dd9a04ea21" (UID: "5b99319e-4fc6-4cfc-939c-14dd9a04ea21"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.259057 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.259086 4959 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.259110 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prvsh\" (UniqueName: \"kubernetes.io/projected/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-kube-api-access-prvsh\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.259120 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.259129 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.259137 4959 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.259144 4959 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b99319e-4fc6-4cfc-939c-14dd9a04ea21-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.435365 4959 generic.go:334] "Generic (PLEG): container finished" podID="5b99319e-4fc6-4cfc-939c-14dd9a04ea21" containerID="8b12ee5c290b974cdbb5d92113026cb92ba6b5dec8c5771a34c09a32d9197b14" exitCode=0 Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.435401 4959 generic.go:334] "Generic (PLEG): container finished" podID="5b99319e-4fc6-4cfc-939c-14dd9a04ea21" containerID="1fd6a48cf687a2cbaea1ddac89b0a3b3b2baf2eac2d793e8a0d86a733826359c" exitCode=2 Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.435408 4959 generic.go:334] "Generic (PLEG): container finished" podID="5b99319e-4fc6-4cfc-939c-14dd9a04ea21" containerID="8d9e92ff000993b1e005b5c4259b07c490aa0a2badf1c746cb1d741de3aca253" exitCode=0 Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.435416 4959 generic.go:334] "Generic (PLEG): container finished" podID="5b99319e-4fc6-4cfc-939c-14dd9a04ea21" containerID="eb81d79bfd2cc50651cd8c8486b6ec8d06cc2e33553e74fda7ad7a680ae04dbb" exitCode=0 Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.435592 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.436353 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b99319e-4fc6-4cfc-939c-14dd9a04ea21","Type":"ContainerDied","Data":"8b12ee5c290b974cdbb5d92113026cb92ba6b5dec8c5771a34c09a32d9197b14"} Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.436398 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b99319e-4fc6-4cfc-939c-14dd9a04ea21","Type":"ContainerDied","Data":"1fd6a48cf687a2cbaea1ddac89b0a3b3b2baf2eac2d793e8a0d86a733826359c"} Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.436410 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b99319e-4fc6-4cfc-939c-14dd9a04ea21","Type":"ContainerDied","Data":"8d9e92ff000993b1e005b5c4259b07c490aa0a2badf1c746cb1d741de3aca253"} Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.436419 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b99319e-4fc6-4cfc-939c-14dd9a04ea21","Type":"ContainerDied","Data":"eb81d79bfd2cc50651cd8c8486b6ec8d06cc2e33553e74fda7ad7a680ae04dbb"} Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.436428 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b99319e-4fc6-4cfc-939c-14dd9a04ea21","Type":"ContainerDied","Data":"22f8c634268c3e3303a3b67d0c33691e87b927ee0823fc174e7eb000dbe06fea"} Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.436447 4959 scope.go:117] "RemoveContainer" containerID="8b12ee5c290b974cdbb5d92113026cb92ba6b5dec8c5771a34c09a32d9197b14" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.472718 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.483315 4959 scope.go:117] "RemoveContainer" containerID="1fd6a48cf687a2cbaea1ddac89b0a3b3b2baf2eac2d793e8a0d86a733826359c" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.491329 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.504536 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:04:22 crc kubenswrapper[4959]: E1007 14:04:22.505082 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b99319e-4fc6-4cfc-939c-14dd9a04ea21" containerName="proxy-httpd" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.505122 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b99319e-4fc6-4cfc-939c-14dd9a04ea21" containerName="proxy-httpd" Oct 07 14:04:22 crc kubenswrapper[4959]: E1007 14:04:22.505146 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b99319e-4fc6-4cfc-939c-14dd9a04ea21" containerName="ceilometer-central-agent" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.505175 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b99319e-4fc6-4cfc-939c-14dd9a04ea21" containerName="ceilometer-central-agent" Oct 07 14:04:22 crc kubenswrapper[4959]: E1007 14:04:22.505191 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b99319e-4fc6-4cfc-939c-14dd9a04ea21" containerName="ceilometer-notification-agent" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.505200 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b99319e-4fc6-4cfc-939c-14dd9a04ea21" containerName="ceilometer-notification-agent" Oct 07 14:04:22 crc kubenswrapper[4959]: E1007 14:04:22.505233 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b99319e-4fc6-4cfc-939c-14dd9a04ea21" containerName="sg-core" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.505242 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b99319e-4fc6-4cfc-939c-14dd9a04ea21" containerName="sg-core" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.505448 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b99319e-4fc6-4cfc-939c-14dd9a04ea21" containerName="sg-core" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.505471 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b99319e-4fc6-4cfc-939c-14dd9a04ea21" containerName="ceilometer-central-agent" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.505500 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b99319e-4fc6-4cfc-939c-14dd9a04ea21" containerName="proxy-httpd" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.505512 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b99319e-4fc6-4cfc-939c-14dd9a04ea21" containerName="ceilometer-notification-agent" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.507461 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.510084 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.514282 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.525431 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.548246 4959 scope.go:117] "RemoveContainer" containerID="8d9e92ff000993b1e005b5c4259b07c490aa0a2badf1c746cb1d741de3aca253" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.620253 4959 scope.go:117] "RemoveContainer" containerID="eb81d79bfd2cc50651cd8c8486b6ec8d06cc2e33553e74fda7ad7a680ae04dbb" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.649801 4959 scope.go:117] "RemoveContainer" containerID="8b12ee5c290b974cdbb5d92113026cb92ba6b5dec8c5771a34c09a32d9197b14" Oct 07 14:04:22 crc kubenswrapper[4959]: E1007 14:04:22.650485 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b12ee5c290b974cdbb5d92113026cb92ba6b5dec8c5771a34c09a32d9197b14\": container with ID starting with 8b12ee5c290b974cdbb5d92113026cb92ba6b5dec8c5771a34c09a32d9197b14 not found: ID does not exist" containerID="8b12ee5c290b974cdbb5d92113026cb92ba6b5dec8c5771a34c09a32d9197b14" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.650525 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b12ee5c290b974cdbb5d92113026cb92ba6b5dec8c5771a34c09a32d9197b14"} err="failed to get container status \"8b12ee5c290b974cdbb5d92113026cb92ba6b5dec8c5771a34c09a32d9197b14\": rpc error: code = NotFound desc = could not find container \"8b12ee5c290b974cdbb5d92113026cb92ba6b5dec8c5771a34c09a32d9197b14\": container with ID starting with 8b12ee5c290b974cdbb5d92113026cb92ba6b5dec8c5771a34c09a32d9197b14 not found: ID does not exist" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.650549 4959 scope.go:117] "RemoveContainer" containerID="1fd6a48cf687a2cbaea1ddac89b0a3b3b2baf2eac2d793e8a0d86a733826359c" Oct 07 14:04:22 crc kubenswrapper[4959]: E1007 14:04:22.650994 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fd6a48cf687a2cbaea1ddac89b0a3b3b2baf2eac2d793e8a0d86a733826359c\": container with ID starting with 1fd6a48cf687a2cbaea1ddac89b0a3b3b2baf2eac2d793e8a0d86a733826359c not found: ID does not exist" containerID="1fd6a48cf687a2cbaea1ddac89b0a3b3b2baf2eac2d793e8a0d86a733826359c" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.651031 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fd6a48cf687a2cbaea1ddac89b0a3b3b2baf2eac2d793e8a0d86a733826359c"} err="failed to get container status \"1fd6a48cf687a2cbaea1ddac89b0a3b3b2baf2eac2d793e8a0d86a733826359c\": rpc error: code = NotFound desc = could not find container \"1fd6a48cf687a2cbaea1ddac89b0a3b3b2baf2eac2d793e8a0d86a733826359c\": container with ID starting with 1fd6a48cf687a2cbaea1ddac89b0a3b3b2baf2eac2d793e8a0d86a733826359c not found: ID does not exist" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.651063 4959 scope.go:117] "RemoveContainer" containerID="8d9e92ff000993b1e005b5c4259b07c490aa0a2badf1c746cb1d741de3aca253" Oct 07 14:04:22 crc kubenswrapper[4959]: E1007 14:04:22.651361 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d9e92ff000993b1e005b5c4259b07c490aa0a2badf1c746cb1d741de3aca253\": container with ID starting with 8d9e92ff000993b1e005b5c4259b07c490aa0a2badf1c746cb1d741de3aca253 not found: ID does not exist" containerID="8d9e92ff000993b1e005b5c4259b07c490aa0a2badf1c746cb1d741de3aca253" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.651395 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d9e92ff000993b1e005b5c4259b07c490aa0a2badf1c746cb1d741de3aca253"} err="failed to get container status \"8d9e92ff000993b1e005b5c4259b07c490aa0a2badf1c746cb1d741de3aca253\": rpc error: code = NotFound desc = could not find container \"8d9e92ff000993b1e005b5c4259b07c490aa0a2badf1c746cb1d741de3aca253\": container with ID starting with 8d9e92ff000993b1e005b5c4259b07c490aa0a2badf1c746cb1d741de3aca253 not found: ID does not exist" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.651410 4959 scope.go:117] "RemoveContainer" containerID="eb81d79bfd2cc50651cd8c8486b6ec8d06cc2e33553e74fda7ad7a680ae04dbb" Oct 07 14:04:22 crc kubenswrapper[4959]: E1007 14:04:22.651793 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb81d79bfd2cc50651cd8c8486b6ec8d06cc2e33553e74fda7ad7a680ae04dbb\": container with ID starting with eb81d79bfd2cc50651cd8c8486b6ec8d06cc2e33553e74fda7ad7a680ae04dbb not found: ID does not exist" containerID="eb81d79bfd2cc50651cd8c8486b6ec8d06cc2e33553e74fda7ad7a680ae04dbb" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.651835 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb81d79bfd2cc50651cd8c8486b6ec8d06cc2e33553e74fda7ad7a680ae04dbb"} err="failed to get container status \"eb81d79bfd2cc50651cd8c8486b6ec8d06cc2e33553e74fda7ad7a680ae04dbb\": rpc error: code = NotFound desc = could not find container \"eb81d79bfd2cc50651cd8c8486b6ec8d06cc2e33553e74fda7ad7a680ae04dbb\": container with ID starting with eb81d79bfd2cc50651cd8c8486b6ec8d06cc2e33553e74fda7ad7a680ae04dbb not found: ID does not exist" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.651862 4959 scope.go:117] "RemoveContainer" containerID="8b12ee5c290b974cdbb5d92113026cb92ba6b5dec8c5771a34c09a32d9197b14" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.652252 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b12ee5c290b974cdbb5d92113026cb92ba6b5dec8c5771a34c09a32d9197b14"} err="failed to get container status \"8b12ee5c290b974cdbb5d92113026cb92ba6b5dec8c5771a34c09a32d9197b14\": rpc error: code = NotFound desc = could not find container \"8b12ee5c290b974cdbb5d92113026cb92ba6b5dec8c5771a34c09a32d9197b14\": container with ID starting with 8b12ee5c290b974cdbb5d92113026cb92ba6b5dec8c5771a34c09a32d9197b14 not found: ID does not exist" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.652270 4959 scope.go:117] "RemoveContainer" containerID="1fd6a48cf687a2cbaea1ddac89b0a3b3b2baf2eac2d793e8a0d86a733826359c" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.653595 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fd6a48cf687a2cbaea1ddac89b0a3b3b2baf2eac2d793e8a0d86a733826359c"} err="failed to get container status \"1fd6a48cf687a2cbaea1ddac89b0a3b3b2baf2eac2d793e8a0d86a733826359c\": rpc error: code = NotFound desc = could not find container \"1fd6a48cf687a2cbaea1ddac89b0a3b3b2baf2eac2d793e8a0d86a733826359c\": container with ID starting with 1fd6a48cf687a2cbaea1ddac89b0a3b3b2baf2eac2d793e8a0d86a733826359c not found: ID does not exist" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.653615 4959 scope.go:117] "RemoveContainer" containerID="8d9e92ff000993b1e005b5c4259b07c490aa0a2badf1c746cb1d741de3aca253" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.653924 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d9e92ff000993b1e005b5c4259b07c490aa0a2badf1c746cb1d741de3aca253"} err="failed to get container status \"8d9e92ff000993b1e005b5c4259b07c490aa0a2badf1c746cb1d741de3aca253\": rpc error: code = NotFound desc = could not find container \"8d9e92ff000993b1e005b5c4259b07c490aa0a2badf1c746cb1d741de3aca253\": container with ID starting with 8d9e92ff000993b1e005b5c4259b07c490aa0a2badf1c746cb1d741de3aca253 not found: ID does not exist" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.653954 4959 scope.go:117] "RemoveContainer" containerID="eb81d79bfd2cc50651cd8c8486b6ec8d06cc2e33553e74fda7ad7a680ae04dbb" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.654270 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb81d79bfd2cc50651cd8c8486b6ec8d06cc2e33553e74fda7ad7a680ae04dbb"} err="failed to get container status \"eb81d79bfd2cc50651cd8c8486b6ec8d06cc2e33553e74fda7ad7a680ae04dbb\": rpc error: code = NotFound desc = could not find container \"eb81d79bfd2cc50651cd8c8486b6ec8d06cc2e33553e74fda7ad7a680ae04dbb\": container with ID starting with eb81d79bfd2cc50651cd8c8486b6ec8d06cc2e33553e74fda7ad7a680ae04dbb not found: ID does not exist" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.654288 4959 scope.go:117] "RemoveContainer" containerID="8b12ee5c290b974cdbb5d92113026cb92ba6b5dec8c5771a34c09a32d9197b14" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.654563 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b12ee5c290b974cdbb5d92113026cb92ba6b5dec8c5771a34c09a32d9197b14"} err="failed to get container status \"8b12ee5c290b974cdbb5d92113026cb92ba6b5dec8c5771a34c09a32d9197b14\": rpc error: code = NotFound desc = could not find container \"8b12ee5c290b974cdbb5d92113026cb92ba6b5dec8c5771a34c09a32d9197b14\": container with ID starting with 8b12ee5c290b974cdbb5d92113026cb92ba6b5dec8c5771a34c09a32d9197b14 not found: ID does not exist" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.654579 4959 scope.go:117] "RemoveContainer" containerID="1fd6a48cf687a2cbaea1ddac89b0a3b3b2baf2eac2d793e8a0d86a733826359c" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.654838 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fd6a48cf687a2cbaea1ddac89b0a3b3b2baf2eac2d793e8a0d86a733826359c"} err="failed to get container status \"1fd6a48cf687a2cbaea1ddac89b0a3b3b2baf2eac2d793e8a0d86a733826359c\": rpc error: code = NotFound desc = could not find container \"1fd6a48cf687a2cbaea1ddac89b0a3b3b2baf2eac2d793e8a0d86a733826359c\": container with ID starting with 1fd6a48cf687a2cbaea1ddac89b0a3b3b2baf2eac2d793e8a0d86a733826359c not found: ID does not exist" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.654858 4959 scope.go:117] "RemoveContainer" containerID="8d9e92ff000993b1e005b5c4259b07c490aa0a2badf1c746cb1d741de3aca253" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.655136 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d9e92ff000993b1e005b5c4259b07c490aa0a2badf1c746cb1d741de3aca253"} err="failed to get container status \"8d9e92ff000993b1e005b5c4259b07c490aa0a2badf1c746cb1d741de3aca253\": rpc error: code = NotFound desc = could not find container \"8d9e92ff000993b1e005b5c4259b07c490aa0a2badf1c746cb1d741de3aca253\": container with ID starting with 8d9e92ff000993b1e005b5c4259b07c490aa0a2badf1c746cb1d741de3aca253 not found: ID does not exist" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.655179 4959 scope.go:117] "RemoveContainer" containerID="eb81d79bfd2cc50651cd8c8486b6ec8d06cc2e33553e74fda7ad7a680ae04dbb" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.655441 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb81d79bfd2cc50651cd8c8486b6ec8d06cc2e33553e74fda7ad7a680ae04dbb"} err="failed to get container status \"eb81d79bfd2cc50651cd8c8486b6ec8d06cc2e33553e74fda7ad7a680ae04dbb\": rpc error: code = NotFound desc = could not find container \"eb81d79bfd2cc50651cd8c8486b6ec8d06cc2e33553e74fda7ad7a680ae04dbb\": container with ID starting with eb81d79bfd2cc50651cd8c8486b6ec8d06cc2e33553e74fda7ad7a680ae04dbb not found: ID does not exist" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.655469 4959 scope.go:117] "RemoveContainer" containerID="8b12ee5c290b974cdbb5d92113026cb92ba6b5dec8c5771a34c09a32d9197b14" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.655686 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b12ee5c290b974cdbb5d92113026cb92ba6b5dec8c5771a34c09a32d9197b14"} err="failed to get container status \"8b12ee5c290b974cdbb5d92113026cb92ba6b5dec8c5771a34c09a32d9197b14\": rpc error: code = NotFound desc = could not find container \"8b12ee5c290b974cdbb5d92113026cb92ba6b5dec8c5771a34c09a32d9197b14\": container with ID starting with 8b12ee5c290b974cdbb5d92113026cb92ba6b5dec8c5771a34c09a32d9197b14 not found: ID does not exist" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.655721 4959 scope.go:117] "RemoveContainer" containerID="1fd6a48cf687a2cbaea1ddac89b0a3b3b2baf2eac2d793e8a0d86a733826359c" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.655983 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fd6a48cf687a2cbaea1ddac89b0a3b3b2baf2eac2d793e8a0d86a733826359c"} err="failed to get container status \"1fd6a48cf687a2cbaea1ddac89b0a3b3b2baf2eac2d793e8a0d86a733826359c\": rpc error: code = NotFound desc = could not find container \"1fd6a48cf687a2cbaea1ddac89b0a3b3b2baf2eac2d793e8a0d86a733826359c\": container with ID starting with 1fd6a48cf687a2cbaea1ddac89b0a3b3b2baf2eac2d793e8a0d86a733826359c not found: ID does not exist" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.656003 4959 scope.go:117] "RemoveContainer" containerID="8d9e92ff000993b1e005b5c4259b07c490aa0a2badf1c746cb1d741de3aca253" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.656617 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d9e92ff000993b1e005b5c4259b07c490aa0a2badf1c746cb1d741de3aca253"} err="failed to get container status \"8d9e92ff000993b1e005b5c4259b07c490aa0a2badf1c746cb1d741de3aca253\": rpc error: code = NotFound desc = could not find container \"8d9e92ff000993b1e005b5c4259b07c490aa0a2badf1c746cb1d741de3aca253\": container with ID starting with 8d9e92ff000993b1e005b5c4259b07c490aa0a2badf1c746cb1d741de3aca253 not found: ID does not exist" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.656637 4959 scope.go:117] "RemoveContainer" containerID="eb81d79bfd2cc50651cd8c8486b6ec8d06cc2e33553e74fda7ad7a680ae04dbb" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.656950 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb81d79bfd2cc50651cd8c8486b6ec8d06cc2e33553e74fda7ad7a680ae04dbb"} err="failed to get container status \"eb81d79bfd2cc50651cd8c8486b6ec8d06cc2e33553e74fda7ad7a680ae04dbb\": rpc error: code = NotFound desc = could not find container \"eb81d79bfd2cc50651cd8c8486b6ec8d06cc2e33553e74fda7ad7a680ae04dbb\": container with ID starting with eb81d79bfd2cc50651cd8c8486b6ec8d06cc2e33553e74fda7ad7a680ae04dbb not found: ID does not exist" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.664844 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b99319e-4fc6-4cfc-939c-14dd9a04ea21" path="/var/lib/kubelet/pods/5b99319e-4fc6-4cfc-939c-14dd9a04ea21/volumes" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.666926 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39d84283-60e5-46de-a5d4-38780edd7e62-scripts\") pod \"ceilometer-0\" (UID: \"39d84283-60e5-46de-a5d4-38780edd7e62\") " pod="openstack/ceilometer-0" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.666961 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wr2f\" (UniqueName: \"kubernetes.io/projected/39d84283-60e5-46de-a5d4-38780edd7e62-kube-api-access-8wr2f\") pod \"ceilometer-0\" (UID: \"39d84283-60e5-46de-a5d4-38780edd7e62\") " pod="openstack/ceilometer-0" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.667003 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39d84283-60e5-46de-a5d4-38780edd7e62-run-httpd\") pod \"ceilometer-0\" (UID: \"39d84283-60e5-46de-a5d4-38780edd7e62\") " pod="openstack/ceilometer-0" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.667034 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39d84283-60e5-46de-a5d4-38780edd7e62-config-data\") pod \"ceilometer-0\" (UID: \"39d84283-60e5-46de-a5d4-38780edd7e62\") " pod="openstack/ceilometer-0" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.667118 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39d84283-60e5-46de-a5d4-38780edd7e62-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"39d84283-60e5-46de-a5d4-38780edd7e62\") " pod="openstack/ceilometer-0" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.667147 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/39d84283-60e5-46de-a5d4-38780edd7e62-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"39d84283-60e5-46de-a5d4-38780edd7e62\") " pod="openstack/ceilometer-0" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.667667 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39d84283-60e5-46de-a5d4-38780edd7e62-log-httpd\") pod \"ceilometer-0\" (UID: \"39d84283-60e5-46de-a5d4-38780edd7e62\") " pod="openstack/ceilometer-0" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.759790 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3e47-account-create-h55ck" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.768880 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39d84283-60e5-46de-a5d4-38780edd7e62-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"39d84283-60e5-46de-a5d4-38780edd7e62\") " pod="openstack/ceilometer-0" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.768934 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/39d84283-60e5-46de-a5d4-38780edd7e62-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"39d84283-60e5-46de-a5d4-38780edd7e62\") " pod="openstack/ceilometer-0" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.768963 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39d84283-60e5-46de-a5d4-38780edd7e62-log-httpd\") pod \"ceilometer-0\" (UID: \"39d84283-60e5-46de-a5d4-38780edd7e62\") " pod="openstack/ceilometer-0" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.769031 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39d84283-60e5-46de-a5d4-38780edd7e62-scripts\") pod \"ceilometer-0\" (UID: \"39d84283-60e5-46de-a5d4-38780edd7e62\") " pod="openstack/ceilometer-0" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.769056 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wr2f\" (UniqueName: \"kubernetes.io/projected/39d84283-60e5-46de-a5d4-38780edd7e62-kube-api-access-8wr2f\") pod \"ceilometer-0\" (UID: \"39d84283-60e5-46de-a5d4-38780edd7e62\") " pod="openstack/ceilometer-0" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.769139 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39d84283-60e5-46de-a5d4-38780edd7e62-run-httpd\") pod \"ceilometer-0\" (UID: \"39d84283-60e5-46de-a5d4-38780edd7e62\") " pod="openstack/ceilometer-0" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.769196 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39d84283-60e5-46de-a5d4-38780edd7e62-config-data\") pod \"ceilometer-0\" (UID: \"39d84283-60e5-46de-a5d4-38780edd7e62\") " pod="openstack/ceilometer-0" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.769956 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39d84283-60e5-46de-a5d4-38780edd7e62-log-httpd\") pod \"ceilometer-0\" (UID: \"39d84283-60e5-46de-a5d4-38780edd7e62\") " pod="openstack/ceilometer-0" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.770125 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39d84283-60e5-46de-a5d4-38780edd7e62-run-httpd\") pod \"ceilometer-0\" (UID: \"39d84283-60e5-46de-a5d4-38780edd7e62\") " pod="openstack/ceilometer-0" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.774367 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39d84283-60e5-46de-a5d4-38780edd7e62-scripts\") pod \"ceilometer-0\" (UID: \"39d84283-60e5-46de-a5d4-38780edd7e62\") " pod="openstack/ceilometer-0" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.774488 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39d84283-60e5-46de-a5d4-38780edd7e62-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"39d84283-60e5-46de-a5d4-38780edd7e62\") " pod="openstack/ceilometer-0" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.776831 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39d84283-60e5-46de-a5d4-38780edd7e62-config-data\") pod \"ceilometer-0\" (UID: \"39d84283-60e5-46de-a5d4-38780edd7e62\") " pod="openstack/ceilometer-0" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.778821 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/39d84283-60e5-46de-a5d4-38780edd7e62-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"39d84283-60e5-46de-a5d4-38780edd7e62\") " pod="openstack/ceilometer-0" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.798394 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wr2f\" (UniqueName: \"kubernetes.io/projected/39d84283-60e5-46de-a5d4-38780edd7e62-kube-api-access-8wr2f\") pod \"ceilometer-0\" (UID: \"39d84283-60e5-46de-a5d4-38780edd7e62\") " pod="openstack/ceilometer-0" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.871145 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksvf5\" (UniqueName: \"kubernetes.io/projected/c9743ad9-caa9-4bc4-97ef-fa52706ca1ae-kube-api-access-ksvf5\") pod \"c9743ad9-caa9-4bc4-97ef-fa52706ca1ae\" (UID: \"c9743ad9-caa9-4bc4-97ef-fa52706ca1ae\") " Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.874584 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9743ad9-caa9-4bc4-97ef-fa52706ca1ae-kube-api-access-ksvf5" (OuterVolumeSpecName: "kube-api-access-ksvf5") pod "c9743ad9-caa9-4bc4-97ef-fa52706ca1ae" (UID: "c9743ad9-caa9-4bc4-97ef-fa52706ca1ae"). InnerVolumeSpecName "kube-api-access-ksvf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.904438 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:04:22 crc kubenswrapper[4959]: I1007 14:04:22.973751 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksvf5\" (UniqueName: \"kubernetes.io/projected/c9743ad9-caa9-4bc4-97ef-fa52706ca1ae-kube-api-access-ksvf5\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:23 crc kubenswrapper[4959]: I1007 14:04:23.327513 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:04:23 crc kubenswrapper[4959]: W1007 14:04:23.338586 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39d84283_60e5_46de_a5d4_38780edd7e62.slice/crio-430459e58b12d7180c615033e8d91d36ec0a53c5a91df3cc2fa6e5839237b052 WatchSource:0}: Error finding container 430459e58b12d7180c615033e8d91d36ec0a53c5a91df3cc2fa6e5839237b052: Status 404 returned error can't find the container with id 430459e58b12d7180c615033e8d91d36ec0a53c5a91df3cc2fa6e5839237b052 Oct 07 14:04:23 crc kubenswrapper[4959]: I1007 14:04:23.446668 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39d84283-60e5-46de-a5d4-38780edd7e62","Type":"ContainerStarted","Data":"430459e58b12d7180c615033e8d91d36ec0a53c5a91df3cc2fa6e5839237b052"} Oct 07 14:04:23 crc kubenswrapper[4959]: I1007 14:04:23.448076 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3e47-account-create-h55ck" event={"ID":"c9743ad9-caa9-4bc4-97ef-fa52706ca1ae","Type":"ContainerDied","Data":"445514157d1e163c9489d671e8c54cc1740606237a4cdf485113a5956ef0f864"} Oct 07 14:04:23 crc kubenswrapper[4959]: I1007 14:04:23.448117 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="445514157d1e163c9489d671e8c54cc1740606237a4cdf485113a5956ef0f864" Oct 07 14:04:23 crc kubenswrapper[4959]: I1007 14:04:23.448175 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3e47-account-create-h55ck" Oct 07 14:04:24 crc kubenswrapper[4959]: I1007 14:04:24.457493 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39d84283-60e5-46de-a5d4-38780edd7e62","Type":"ContainerStarted","Data":"3f2dadad232bc872a1e8c11f4cdc935a61f204b6546cbe1dd82d3619ccd1d9c4"} Oct 07 14:04:25 crc kubenswrapper[4959]: I1007 14:04:25.466375 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39d84283-60e5-46de-a5d4-38780edd7e62","Type":"ContainerStarted","Data":"5147cc24a64ee80837f4aee713ad2ff586d2da038035a705298febf260a0b11f"} Oct 07 14:04:25 crc kubenswrapper[4959]: I1007 14:04:25.854177 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:04:26 crc kubenswrapper[4959]: I1007 14:04:26.476226 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39d84283-60e5-46de-a5d4-38780edd7e62","Type":"ContainerStarted","Data":"34169a80cb0e921bf0c34f48993b822cfcab288502ca522d73c8bd9e645b8ccd"} Oct 07 14:04:26 crc kubenswrapper[4959]: I1007 14:04:26.489227 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-744bf8f46d-6twnn" podUID="a588fbdc-fd65-49f6-ab59-4901c7ca690a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Oct 07 14:04:26 crc kubenswrapper[4959]: I1007 14:04:26.489328 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:04:28 crc kubenswrapper[4959]: I1007 14:04:28.496655 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39d84283-60e5-46de-a5d4-38780edd7e62","Type":"ContainerStarted","Data":"fdadb49cec4d99c1eb89e6e3d1c5a86b29101d2234f6c48d078fb61be8954f2a"} Oct 07 14:04:28 crc kubenswrapper[4959]: I1007 14:04:28.497230 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 07 14:04:28 crc kubenswrapper[4959]: I1007 14:04:28.496799 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="39d84283-60e5-46de-a5d4-38780edd7e62" containerName="proxy-httpd" containerID="cri-o://fdadb49cec4d99c1eb89e6e3d1c5a86b29101d2234f6c48d078fb61be8954f2a" gracePeriod=30 Oct 07 14:04:28 crc kubenswrapper[4959]: I1007 14:04:28.496766 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="39d84283-60e5-46de-a5d4-38780edd7e62" containerName="ceilometer-central-agent" containerID="cri-o://3f2dadad232bc872a1e8c11f4cdc935a61f204b6546cbe1dd82d3619ccd1d9c4" gracePeriod=30 Oct 07 14:04:28 crc kubenswrapper[4959]: I1007 14:04:28.496887 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="39d84283-60e5-46de-a5d4-38780edd7e62" containerName="ceilometer-notification-agent" containerID="cri-o://5147cc24a64ee80837f4aee713ad2ff586d2da038035a705298febf260a0b11f" gracePeriod=30 Oct 07 14:04:28 crc kubenswrapper[4959]: I1007 14:04:28.496943 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="39d84283-60e5-46de-a5d4-38780edd7e62" containerName="sg-core" containerID="cri-o://34169a80cb0e921bf0c34f48993b822cfcab288502ca522d73c8bd9e645b8ccd" gracePeriod=30 Oct 07 14:04:28 crc kubenswrapper[4959]: I1007 14:04:28.520062 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.421368547 podStartE2EDuration="6.520042273s" podCreationTimestamp="2025-10-07 14:04:22 +0000 UTC" firstStartedPulling="2025-10-07 14:04:23.341958142 +0000 UTC m=+1145.425362467" lastFinishedPulling="2025-10-07 14:04:27.440631858 +0000 UTC m=+1149.524036193" observedRunningTime="2025-10-07 14:04:28.517380107 +0000 UTC m=+1150.600784462" watchObservedRunningTime="2025-10-07 14:04:28.520042273 +0000 UTC m=+1150.603446598" Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.506846 4959 generic.go:334] "Generic (PLEG): container finished" podID="39d84283-60e5-46de-a5d4-38780edd7e62" containerID="fdadb49cec4d99c1eb89e6e3d1c5a86b29101d2234f6c48d078fb61be8954f2a" exitCode=0 Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.507202 4959 generic.go:334] "Generic (PLEG): container finished" podID="39d84283-60e5-46de-a5d4-38780edd7e62" containerID="34169a80cb0e921bf0c34f48993b822cfcab288502ca522d73c8bd9e645b8ccd" exitCode=2 Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.507214 4959 generic.go:334] "Generic (PLEG): container finished" podID="39d84283-60e5-46de-a5d4-38780edd7e62" containerID="5147cc24a64ee80837f4aee713ad2ff586d2da038035a705298febf260a0b11f" exitCode=0 Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.506911 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39d84283-60e5-46de-a5d4-38780edd7e62","Type":"ContainerDied","Data":"fdadb49cec4d99c1eb89e6e3d1c5a86b29101d2234f6c48d078fb61be8954f2a"} Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.507243 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39d84283-60e5-46de-a5d4-38780edd7e62","Type":"ContainerDied","Data":"34169a80cb0e921bf0c34f48993b822cfcab288502ca522d73c8bd9e645b8ccd"} Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.507255 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39d84283-60e5-46de-a5d4-38780edd7e62","Type":"ContainerDied","Data":"5147cc24a64ee80837f4aee713ad2ff586d2da038035a705298febf260a0b11f"} Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.807797 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-1479-account-create-n9c4j"] Oct 07 14:04:29 crc kubenswrapper[4959]: E1007 14:04:29.808374 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9743ad9-caa9-4bc4-97ef-fa52706ca1ae" containerName="mariadb-account-create" Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.808401 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9743ad9-caa9-4bc4-97ef-fa52706ca1ae" containerName="mariadb-account-create" Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.809872 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9743ad9-caa9-4bc4-97ef-fa52706ca1ae" containerName="mariadb-account-create" Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.810608 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-1479-account-create-n9c4j" Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.812385 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.817174 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-1479-account-create-n9c4j"] Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.853719 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.994519 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-79f7-account-create-7n7bb"] Oct 07 14:04:29 crc kubenswrapper[4959]: E1007 14:04:29.994881 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39d84283-60e5-46de-a5d4-38780edd7e62" containerName="ceilometer-central-agent" Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.994894 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="39d84283-60e5-46de-a5d4-38780edd7e62" containerName="ceilometer-central-agent" Oct 07 14:04:29 crc kubenswrapper[4959]: E1007 14:04:29.994907 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39d84283-60e5-46de-a5d4-38780edd7e62" containerName="sg-core" Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.994913 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="39d84283-60e5-46de-a5d4-38780edd7e62" containerName="sg-core" Oct 07 14:04:29 crc kubenswrapper[4959]: E1007 14:04:29.994932 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39d84283-60e5-46de-a5d4-38780edd7e62" containerName="proxy-httpd" Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.994938 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="39d84283-60e5-46de-a5d4-38780edd7e62" containerName="proxy-httpd" Oct 07 14:04:29 crc kubenswrapper[4959]: E1007 14:04:29.994950 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39d84283-60e5-46de-a5d4-38780edd7e62" containerName="ceilometer-notification-agent" Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.994955 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="39d84283-60e5-46de-a5d4-38780edd7e62" containerName="ceilometer-notification-agent" Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.995145 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="39d84283-60e5-46de-a5d4-38780edd7e62" containerName="sg-core" Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.995159 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="39d84283-60e5-46de-a5d4-38780edd7e62" containerName="proxy-httpd" Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.995183 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="39d84283-60e5-46de-a5d4-38780edd7e62" containerName="ceilometer-notification-agent" Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.995192 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="39d84283-60e5-46de-a5d4-38780edd7e62" containerName="ceilometer-central-agent" Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.995675 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-79f7-account-create-7n7bb" Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.995808 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39d84283-60e5-46de-a5d4-38780edd7e62-scripts\") pod \"39d84283-60e5-46de-a5d4-38780edd7e62\" (UID: \"39d84283-60e5-46de-a5d4-38780edd7e62\") " Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.996006 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39d84283-60e5-46de-a5d4-38780edd7e62-run-httpd\") pod \"39d84283-60e5-46de-a5d4-38780edd7e62\" (UID: \"39d84283-60e5-46de-a5d4-38780edd7e62\") " Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.996054 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39d84283-60e5-46de-a5d4-38780edd7e62-combined-ca-bundle\") pod \"39d84283-60e5-46de-a5d4-38780edd7e62\" (UID: \"39d84283-60e5-46de-a5d4-38780edd7e62\") " Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.996084 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wr2f\" (UniqueName: \"kubernetes.io/projected/39d84283-60e5-46de-a5d4-38780edd7e62-kube-api-access-8wr2f\") pod \"39d84283-60e5-46de-a5d4-38780edd7e62\" (UID: \"39d84283-60e5-46de-a5d4-38780edd7e62\") " Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.996155 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39d84283-60e5-46de-a5d4-38780edd7e62-log-httpd\") pod \"39d84283-60e5-46de-a5d4-38780edd7e62\" (UID: \"39d84283-60e5-46de-a5d4-38780edd7e62\") " Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.996183 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/39d84283-60e5-46de-a5d4-38780edd7e62-sg-core-conf-yaml\") pod \"39d84283-60e5-46de-a5d4-38780edd7e62\" (UID: \"39d84283-60e5-46de-a5d4-38780edd7e62\") " Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.996228 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39d84283-60e5-46de-a5d4-38780edd7e62-config-data\") pod \"39d84283-60e5-46de-a5d4-38780edd7e62\" (UID: \"39d84283-60e5-46de-a5d4-38780edd7e62\") " Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.996549 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpkbg\" (UniqueName: \"kubernetes.io/projected/63e77afb-d7b9-46a0-8f3c-e12d8157812d-kube-api-access-hpkbg\") pod \"nova-cell0-1479-account-create-n9c4j\" (UID: \"63e77afb-d7b9-46a0-8f3c-e12d8157812d\") " pod="openstack/nova-cell0-1479-account-create-n9c4j" Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.996826 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39d84283-60e5-46de-a5d4-38780edd7e62-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "39d84283-60e5-46de-a5d4-38780edd7e62" (UID: "39d84283-60e5-46de-a5d4-38780edd7e62"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.997449 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39d84283-60e5-46de-a5d4-38780edd7e62-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "39d84283-60e5-46de-a5d4-38780edd7e62" (UID: "39d84283-60e5-46de-a5d4-38780edd7e62"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:04:29 crc kubenswrapper[4959]: I1007 14:04:29.997967 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.002753 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39d84283-60e5-46de-a5d4-38780edd7e62-scripts" (OuterVolumeSpecName: "scripts") pod "39d84283-60e5-46de-a5d4-38780edd7e62" (UID: "39d84283-60e5-46de-a5d4-38780edd7e62"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.004760 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39d84283-60e5-46de-a5d4-38780edd7e62-kube-api-access-8wr2f" (OuterVolumeSpecName: "kube-api-access-8wr2f") pod "39d84283-60e5-46de-a5d4-38780edd7e62" (UID: "39d84283-60e5-46de-a5d4-38780edd7e62"). InnerVolumeSpecName "kube-api-access-8wr2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.006935 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-79f7-account-create-7n7bb"] Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.041390 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39d84283-60e5-46de-a5d4-38780edd7e62-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "39d84283-60e5-46de-a5d4-38780edd7e62" (UID: "39d84283-60e5-46de-a5d4-38780edd7e62"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.083008 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39d84283-60e5-46de-a5d4-38780edd7e62-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "39d84283-60e5-46de-a5d4-38780edd7e62" (UID: "39d84283-60e5-46de-a5d4-38780edd7e62"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.098026 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpkbg\" (UniqueName: \"kubernetes.io/projected/63e77afb-d7b9-46a0-8f3c-e12d8157812d-kube-api-access-hpkbg\") pod \"nova-cell0-1479-account-create-n9c4j\" (UID: \"63e77afb-d7b9-46a0-8f3c-e12d8157812d\") " pod="openstack/nova-cell0-1479-account-create-n9c4j" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.098151 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9q4kb\" (UniqueName: \"kubernetes.io/projected/80610e88-4157-4a37-bbe6-97d422d5f97e-kube-api-access-9q4kb\") pod \"nova-cell1-79f7-account-create-7n7bb\" (UID: \"80610e88-4157-4a37-bbe6-97d422d5f97e\") " pod="openstack/nova-cell1-79f7-account-create-7n7bb" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.098230 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39d84283-60e5-46de-a5d4-38780edd7e62-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.098247 4959 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39d84283-60e5-46de-a5d4-38780edd7e62-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.098284 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39d84283-60e5-46de-a5d4-38780edd7e62-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.098296 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wr2f\" (UniqueName: \"kubernetes.io/projected/39d84283-60e5-46de-a5d4-38780edd7e62-kube-api-access-8wr2f\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.098305 4959 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39d84283-60e5-46de-a5d4-38780edd7e62-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.098312 4959 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/39d84283-60e5-46de-a5d4-38780edd7e62-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.113550 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpkbg\" (UniqueName: \"kubernetes.io/projected/63e77afb-d7b9-46a0-8f3c-e12d8157812d-kube-api-access-hpkbg\") pod \"nova-cell0-1479-account-create-n9c4j\" (UID: \"63e77afb-d7b9-46a0-8f3c-e12d8157812d\") " pod="openstack/nova-cell0-1479-account-create-n9c4j" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.120007 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39d84283-60e5-46de-a5d4-38780edd7e62-config-data" (OuterVolumeSpecName: "config-data") pod "39d84283-60e5-46de-a5d4-38780edd7e62" (UID: "39d84283-60e5-46de-a5d4-38780edd7e62"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.162742 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-1479-account-create-n9c4j" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.200058 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9q4kb\" (UniqueName: \"kubernetes.io/projected/80610e88-4157-4a37-bbe6-97d422d5f97e-kube-api-access-9q4kb\") pod \"nova-cell1-79f7-account-create-7n7bb\" (UID: \"80610e88-4157-4a37-bbe6-97d422d5f97e\") " pod="openstack/nova-cell1-79f7-account-create-7n7bb" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.200399 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39d84283-60e5-46de-a5d4-38780edd7e62-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.217583 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9q4kb\" (UniqueName: \"kubernetes.io/projected/80610e88-4157-4a37-bbe6-97d422d5f97e-kube-api-access-9q4kb\") pod \"nova-cell1-79f7-account-create-7n7bb\" (UID: \"80610e88-4157-4a37-bbe6-97d422d5f97e\") " pod="openstack/nova-cell1-79f7-account-create-7n7bb" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.436845 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-79f7-account-create-7n7bb" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.517417 4959 generic.go:334] "Generic (PLEG): container finished" podID="a588fbdc-fd65-49f6-ab59-4901c7ca690a" containerID="0b8d35aaa990504cd841339a7496639b4a5b31cf2af6d3b62946d1c9d29fa02f" exitCode=137 Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.517471 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-744bf8f46d-6twnn" event={"ID":"a588fbdc-fd65-49f6-ab59-4901c7ca690a","Type":"ContainerDied","Data":"0b8d35aaa990504cd841339a7496639b4a5b31cf2af6d3b62946d1c9d29fa02f"} Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.519749 4959 generic.go:334] "Generic (PLEG): container finished" podID="39d84283-60e5-46de-a5d4-38780edd7e62" containerID="3f2dadad232bc872a1e8c11f4cdc935a61f204b6546cbe1dd82d3619ccd1d9c4" exitCode=0 Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.519767 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39d84283-60e5-46de-a5d4-38780edd7e62","Type":"ContainerDied","Data":"3f2dadad232bc872a1e8c11f4cdc935a61f204b6546cbe1dd82d3619ccd1d9c4"} Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.519783 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39d84283-60e5-46de-a5d4-38780edd7e62","Type":"ContainerDied","Data":"430459e58b12d7180c615033e8d91d36ec0a53c5a91df3cc2fa6e5839237b052"} Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.519799 4959 scope.go:117] "RemoveContainer" containerID="fdadb49cec4d99c1eb89e6e3d1c5a86b29101d2234f6c48d078fb61be8954f2a" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.519908 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.547606 4959 scope.go:117] "RemoveContainer" containerID="34169a80cb0e921bf0c34f48993b822cfcab288502ca522d73c8bd9e645b8ccd" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.560187 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.570894 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.579211 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.581666 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.585683 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.586213 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.586892 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.589794 4959 scope.go:117] "RemoveContainer" containerID="5147cc24a64ee80837f4aee713ad2ff586d2da038035a705298febf260a0b11f" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.631470 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-1479-account-create-n9c4j"] Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.644341 4959 scope.go:117] "RemoveContainer" containerID="3f2dadad232bc872a1e8c11f4cdc935a61f204b6546cbe1dd82d3619ccd1d9c4" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.685718 4959 scope.go:117] "RemoveContainer" containerID="fdadb49cec4d99c1eb89e6e3d1c5a86b29101d2234f6c48d078fb61be8954f2a" Oct 07 14:04:30 crc kubenswrapper[4959]: E1007 14:04:30.686237 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdadb49cec4d99c1eb89e6e3d1c5a86b29101d2234f6c48d078fb61be8954f2a\": container with ID starting with fdadb49cec4d99c1eb89e6e3d1c5a86b29101d2234f6c48d078fb61be8954f2a not found: ID does not exist" containerID="fdadb49cec4d99c1eb89e6e3d1c5a86b29101d2234f6c48d078fb61be8954f2a" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.686287 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdadb49cec4d99c1eb89e6e3d1c5a86b29101d2234f6c48d078fb61be8954f2a"} err="failed to get container status \"fdadb49cec4d99c1eb89e6e3d1c5a86b29101d2234f6c48d078fb61be8954f2a\": rpc error: code = NotFound desc = could not find container \"fdadb49cec4d99c1eb89e6e3d1c5a86b29101d2234f6c48d078fb61be8954f2a\": container with ID starting with fdadb49cec4d99c1eb89e6e3d1c5a86b29101d2234f6c48d078fb61be8954f2a not found: ID does not exist" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.686318 4959 scope.go:117] "RemoveContainer" containerID="34169a80cb0e921bf0c34f48993b822cfcab288502ca522d73c8bd9e645b8ccd" Oct 07 14:04:30 crc kubenswrapper[4959]: E1007 14:04:30.686526 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34169a80cb0e921bf0c34f48993b822cfcab288502ca522d73c8bd9e645b8ccd\": container with ID starting with 34169a80cb0e921bf0c34f48993b822cfcab288502ca522d73c8bd9e645b8ccd not found: ID does not exist" containerID="34169a80cb0e921bf0c34f48993b822cfcab288502ca522d73c8bd9e645b8ccd" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.686549 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34169a80cb0e921bf0c34f48993b822cfcab288502ca522d73c8bd9e645b8ccd"} err="failed to get container status \"34169a80cb0e921bf0c34f48993b822cfcab288502ca522d73c8bd9e645b8ccd\": rpc error: code = NotFound desc = could not find container \"34169a80cb0e921bf0c34f48993b822cfcab288502ca522d73c8bd9e645b8ccd\": container with ID starting with 34169a80cb0e921bf0c34f48993b822cfcab288502ca522d73c8bd9e645b8ccd not found: ID does not exist" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.686567 4959 scope.go:117] "RemoveContainer" containerID="5147cc24a64ee80837f4aee713ad2ff586d2da038035a705298febf260a0b11f" Oct 07 14:04:30 crc kubenswrapper[4959]: E1007 14:04:30.687180 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5147cc24a64ee80837f4aee713ad2ff586d2da038035a705298febf260a0b11f\": container with ID starting with 5147cc24a64ee80837f4aee713ad2ff586d2da038035a705298febf260a0b11f not found: ID does not exist" containerID="5147cc24a64ee80837f4aee713ad2ff586d2da038035a705298febf260a0b11f" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.687216 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5147cc24a64ee80837f4aee713ad2ff586d2da038035a705298febf260a0b11f"} err="failed to get container status \"5147cc24a64ee80837f4aee713ad2ff586d2da038035a705298febf260a0b11f\": rpc error: code = NotFound desc = could not find container \"5147cc24a64ee80837f4aee713ad2ff586d2da038035a705298febf260a0b11f\": container with ID starting with 5147cc24a64ee80837f4aee713ad2ff586d2da038035a705298febf260a0b11f not found: ID does not exist" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.687245 4959 scope.go:117] "RemoveContainer" containerID="3f2dadad232bc872a1e8c11f4cdc935a61f204b6546cbe1dd82d3619ccd1d9c4" Oct 07 14:04:30 crc kubenswrapper[4959]: E1007 14:04:30.688466 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f2dadad232bc872a1e8c11f4cdc935a61f204b6546cbe1dd82d3619ccd1d9c4\": container with ID starting with 3f2dadad232bc872a1e8c11f4cdc935a61f204b6546cbe1dd82d3619ccd1d9c4 not found: ID does not exist" containerID="3f2dadad232bc872a1e8c11f4cdc935a61f204b6546cbe1dd82d3619ccd1d9c4" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.688497 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f2dadad232bc872a1e8c11f4cdc935a61f204b6546cbe1dd82d3619ccd1d9c4"} err="failed to get container status \"3f2dadad232bc872a1e8c11f4cdc935a61f204b6546cbe1dd82d3619ccd1d9c4\": rpc error: code = NotFound desc = could not find container \"3f2dadad232bc872a1e8c11f4cdc935a61f204b6546cbe1dd82d3619ccd1d9c4\": container with ID starting with 3f2dadad232bc872a1e8c11f4cdc935a61f204b6546cbe1dd82d3619ccd1d9c4 not found: ID does not exist" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.689120 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39d84283-60e5-46de-a5d4-38780edd7e62" path="/var/lib/kubelet/pods/39d84283-60e5-46de-a5d4-38780edd7e62/volumes" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.706718 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.707501 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f0b3a16-13a7-42ab-82a6-f2171da788d3-config-data\") pod \"ceilometer-0\" (UID: \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\") " pod="openstack/ceilometer-0" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.707542 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f0b3a16-13a7-42ab-82a6-f2171da788d3-log-httpd\") pod \"ceilometer-0\" (UID: \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\") " pod="openstack/ceilometer-0" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.707566 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f954n\" (UniqueName: \"kubernetes.io/projected/2f0b3a16-13a7-42ab-82a6-f2171da788d3-kube-api-access-f954n\") pod \"ceilometer-0\" (UID: \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\") " pod="openstack/ceilometer-0" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.707727 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f0b3a16-13a7-42ab-82a6-f2171da788d3-run-httpd\") pod \"ceilometer-0\" (UID: \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\") " pod="openstack/ceilometer-0" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.707763 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f0b3a16-13a7-42ab-82a6-f2171da788d3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\") " pod="openstack/ceilometer-0" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.707781 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f0b3a16-13a7-42ab-82a6-f2171da788d3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\") " pod="openstack/ceilometer-0" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.707803 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f0b3a16-13a7-42ab-82a6-f2171da788d3-scripts\") pod \"ceilometer-0\" (UID: \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\") " pod="openstack/ceilometer-0" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.808891 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a588fbdc-fd65-49f6-ab59-4901c7ca690a-config-data\") pod \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\" (UID: \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\") " Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.809238 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a588fbdc-fd65-49f6-ab59-4901c7ca690a-combined-ca-bundle\") pod \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\" (UID: \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\") " Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.809263 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a588fbdc-fd65-49f6-ab59-4901c7ca690a-scripts\") pod \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\" (UID: \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\") " Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.809404 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a588fbdc-fd65-49f6-ab59-4901c7ca690a-horizon-tls-certs\") pod \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\" (UID: \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\") " Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.809523 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9nlx\" (UniqueName: \"kubernetes.io/projected/a588fbdc-fd65-49f6-ab59-4901c7ca690a-kube-api-access-x9nlx\") pod \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\" (UID: \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\") " Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.809565 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a588fbdc-fd65-49f6-ab59-4901c7ca690a-horizon-secret-key\") pod \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\" (UID: \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\") " Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.809587 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a588fbdc-fd65-49f6-ab59-4901c7ca690a-logs\") pod \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\" (UID: \"a588fbdc-fd65-49f6-ab59-4901c7ca690a\") " Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.809852 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f0b3a16-13a7-42ab-82a6-f2171da788d3-run-httpd\") pod \"ceilometer-0\" (UID: \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\") " pod="openstack/ceilometer-0" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.809887 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f0b3a16-13a7-42ab-82a6-f2171da788d3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\") " pod="openstack/ceilometer-0" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.809911 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f0b3a16-13a7-42ab-82a6-f2171da788d3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\") " pod="openstack/ceilometer-0" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.809932 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f0b3a16-13a7-42ab-82a6-f2171da788d3-scripts\") pod \"ceilometer-0\" (UID: \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\") " pod="openstack/ceilometer-0" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.809997 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f0b3a16-13a7-42ab-82a6-f2171da788d3-config-data\") pod \"ceilometer-0\" (UID: \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\") " pod="openstack/ceilometer-0" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.810019 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f0b3a16-13a7-42ab-82a6-f2171da788d3-log-httpd\") pod \"ceilometer-0\" (UID: \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\") " pod="openstack/ceilometer-0" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.810037 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f954n\" (UniqueName: \"kubernetes.io/projected/2f0b3a16-13a7-42ab-82a6-f2171da788d3-kube-api-access-f954n\") pod \"ceilometer-0\" (UID: \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\") " pod="openstack/ceilometer-0" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.811388 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a588fbdc-fd65-49f6-ab59-4901c7ca690a-logs" (OuterVolumeSpecName: "logs") pod "a588fbdc-fd65-49f6-ab59-4901c7ca690a" (UID: "a588fbdc-fd65-49f6-ab59-4901c7ca690a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.811765 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f0b3a16-13a7-42ab-82a6-f2171da788d3-log-httpd\") pod \"ceilometer-0\" (UID: \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\") " pod="openstack/ceilometer-0" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.811820 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f0b3a16-13a7-42ab-82a6-f2171da788d3-run-httpd\") pod \"ceilometer-0\" (UID: \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\") " pod="openstack/ceilometer-0" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.825947 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f0b3a16-13a7-42ab-82a6-f2171da788d3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\") " pod="openstack/ceilometer-0" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.828226 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f0b3a16-13a7-42ab-82a6-f2171da788d3-scripts\") pod \"ceilometer-0\" (UID: \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\") " pod="openstack/ceilometer-0" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.828699 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f954n\" (UniqueName: \"kubernetes.io/projected/2f0b3a16-13a7-42ab-82a6-f2171da788d3-kube-api-access-f954n\") pod \"ceilometer-0\" (UID: \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\") " pod="openstack/ceilometer-0" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.828712 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f0b3a16-13a7-42ab-82a6-f2171da788d3-config-data\") pod \"ceilometer-0\" (UID: \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\") " pod="openstack/ceilometer-0" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.829198 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a588fbdc-fd65-49f6-ab59-4901c7ca690a-kube-api-access-x9nlx" (OuterVolumeSpecName: "kube-api-access-x9nlx") pod "a588fbdc-fd65-49f6-ab59-4901c7ca690a" (UID: "a588fbdc-fd65-49f6-ab59-4901c7ca690a"). InnerVolumeSpecName "kube-api-access-x9nlx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.829500 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f0b3a16-13a7-42ab-82a6-f2171da788d3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\") " pod="openstack/ceilometer-0" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.833498 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a588fbdc-fd65-49f6-ab59-4901c7ca690a-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "a588fbdc-fd65-49f6-ab59-4901c7ca690a" (UID: "a588fbdc-fd65-49f6-ab59-4901c7ca690a"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.836962 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a588fbdc-fd65-49f6-ab59-4901c7ca690a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a588fbdc-fd65-49f6-ab59-4901c7ca690a" (UID: "a588fbdc-fd65-49f6-ab59-4901c7ca690a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.839168 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a588fbdc-fd65-49f6-ab59-4901c7ca690a-scripts" (OuterVolumeSpecName: "scripts") pod "a588fbdc-fd65-49f6-ab59-4901c7ca690a" (UID: "a588fbdc-fd65-49f6-ab59-4901c7ca690a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.846397 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a588fbdc-fd65-49f6-ab59-4901c7ca690a-config-data" (OuterVolumeSpecName: "config-data") pod "a588fbdc-fd65-49f6-ab59-4901c7ca690a" (UID: "a588fbdc-fd65-49f6-ab59-4901c7ca690a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.858309 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a588fbdc-fd65-49f6-ab59-4901c7ca690a-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "a588fbdc-fd65-49f6-ab59-4901c7ca690a" (UID: "a588fbdc-fd65-49f6-ab59-4901c7ca690a"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.911279 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9nlx\" (UniqueName: \"kubernetes.io/projected/a588fbdc-fd65-49f6-ab59-4901c7ca690a-kube-api-access-x9nlx\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.911310 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a588fbdc-fd65-49f6-ab59-4901c7ca690a-logs\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.911360 4959 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a588fbdc-fd65-49f6-ab59-4901c7ca690a-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.911388 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a588fbdc-fd65-49f6-ab59-4901c7ca690a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.911396 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a588fbdc-fd65-49f6-ab59-4901c7ca690a-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.911405 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a588fbdc-fd65-49f6-ab59-4901c7ca690a-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.911412 4959 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a588fbdc-fd65-49f6-ab59-4901c7ca690a-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.911546 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:04:30 crc kubenswrapper[4959]: I1007 14:04:30.959857 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-79f7-account-create-7n7bb"] Oct 07 14:04:30 crc kubenswrapper[4959]: W1007 14:04:30.968483 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod80610e88_4157_4a37_bbe6_97d422d5f97e.slice/crio-a5c73a999de63e8813ce9fd5cd0ee2995e6189dfa72b1fa575de0cf9f385fe65 WatchSource:0}: Error finding container a5c73a999de63e8813ce9fd5cd0ee2995e6189dfa72b1fa575de0cf9f385fe65: Status 404 returned error can't find the container with id a5c73a999de63e8813ce9fd5cd0ee2995e6189dfa72b1fa575de0cf9f385fe65 Oct 07 14:04:31 crc kubenswrapper[4959]: I1007 14:04:31.325521 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:04:31 crc kubenswrapper[4959]: W1007 14:04:31.332497 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f0b3a16_13a7_42ab_82a6_f2171da788d3.slice/crio-1db2f62deebe35c37375ed2df0ef70dd7086d3c1c867ed3707e39b2a2e8d36a6 WatchSource:0}: Error finding container 1db2f62deebe35c37375ed2df0ef70dd7086d3c1c867ed3707e39b2a2e8d36a6: Status 404 returned error can't find the container with id 1db2f62deebe35c37375ed2df0ef70dd7086d3c1c867ed3707e39b2a2e8d36a6 Oct 07 14:04:31 crc kubenswrapper[4959]: I1007 14:04:31.534971 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-744bf8f46d-6twnn" event={"ID":"a588fbdc-fd65-49f6-ab59-4901c7ca690a","Type":"ContainerDied","Data":"0e8c421e3ea812af0ca6b26318aada327d562316382b23587cf36edf6e88d253"} Oct 07 14:04:31 crc kubenswrapper[4959]: I1007 14:04:31.535007 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-744bf8f46d-6twnn" Oct 07 14:04:31 crc kubenswrapper[4959]: I1007 14:04:31.535358 4959 scope.go:117] "RemoveContainer" containerID="76d664db4d06017fdcd093f711c6c672a5feacc3d610e94a71a01f0793720247" Oct 07 14:04:31 crc kubenswrapper[4959]: I1007 14:04:31.541439 4959 generic.go:334] "Generic (PLEG): container finished" podID="80610e88-4157-4a37-bbe6-97d422d5f97e" containerID="4ce2cc6d5f2da117ae8b950eb63cbd16aab79c2ee3381bed94410cb3b7149e32" exitCode=0 Oct 07 14:04:31 crc kubenswrapper[4959]: I1007 14:04:31.541520 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-79f7-account-create-7n7bb" event={"ID":"80610e88-4157-4a37-bbe6-97d422d5f97e","Type":"ContainerDied","Data":"4ce2cc6d5f2da117ae8b950eb63cbd16aab79c2ee3381bed94410cb3b7149e32"} Oct 07 14:04:31 crc kubenswrapper[4959]: I1007 14:04:31.541548 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-79f7-account-create-7n7bb" event={"ID":"80610e88-4157-4a37-bbe6-97d422d5f97e","Type":"ContainerStarted","Data":"a5c73a999de63e8813ce9fd5cd0ee2995e6189dfa72b1fa575de0cf9f385fe65"} Oct 07 14:04:31 crc kubenswrapper[4959]: I1007 14:04:31.544957 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f0b3a16-13a7-42ab-82a6-f2171da788d3","Type":"ContainerStarted","Data":"1db2f62deebe35c37375ed2df0ef70dd7086d3c1c867ed3707e39b2a2e8d36a6"} Oct 07 14:04:31 crc kubenswrapper[4959]: I1007 14:04:31.546344 4959 generic.go:334] "Generic (PLEG): container finished" podID="63e77afb-d7b9-46a0-8f3c-e12d8157812d" containerID="5416daa5ada46ac3981c298e85ab9457898991c449ad80b24295028e6d00fe48" exitCode=0 Oct 07 14:04:31 crc kubenswrapper[4959]: I1007 14:04:31.546381 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-1479-account-create-n9c4j" event={"ID":"63e77afb-d7b9-46a0-8f3c-e12d8157812d","Type":"ContainerDied","Data":"5416daa5ada46ac3981c298e85ab9457898991c449ad80b24295028e6d00fe48"} Oct 07 14:04:31 crc kubenswrapper[4959]: I1007 14:04:31.546410 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-1479-account-create-n9c4j" event={"ID":"63e77afb-d7b9-46a0-8f3c-e12d8157812d","Type":"ContainerStarted","Data":"acdefb3260178dac638f860632af6e11a8698d29c85b46a8bae3f098f90cc003"} Oct 07 14:04:31 crc kubenswrapper[4959]: I1007 14:04:31.579056 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-744bf8f46d-6twnn"] Oct 07 14:04:31 crc kubenswrapper[4959]: I1007 14:04:31.589274 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-744bf8f46d-6twnn"] Oct 07 14:04:31 crc kubenswrapper[4959]: I1007 14:04:31.706868 4959 scope.go:117] "RemoveContainer" containerID="0b8d35aaa990504cd841339a7496639b4a5b31cf2af6d3b62946d1c9d29fa02f" Oct 07 14:04:32 crc kubenswrapper[4959]: I1007 14:04:32.556597 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f0b3a16-13a7-42ab-82a6-f2171da788d3","Type":"ContainerStarted","Data":"5a3b72ba50e34ac0f36f43c40f436380ec0feedb7647c4127fdbaa9f05031e00"} Oct 07 14:04:32 crc kubenswrapper[4959]: I1007 14:04:32.666783 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a588fbdc-fd65-49f6-ab59-4901c7ca690a" path="/var/lib/kubelet/pods/a588fbdc-fd65-49f6-ab59-4901c7ca690a/volumes" Oct 07 14:04:32 crc kubenswrapper[4959]: I1007 14:04:32.967627 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-1479-account-create-n9c4j" Oct 07 14:04:32 crc kubenswrapper[4959]: I1007 14:04:32.972630 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-79f7-account-create-7n7bb" Oct 07 14:04:33 crc kubenswrapper[4959]: I1007 14:04:33.149460 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9q4kb\" (UniqueName: \"kubernetes.io/projected/80610e88-4157-4a37-bbe6-97d422d5f97e-kube-api-access-9q4kb\") pod \"80610e88-4157-4a37-bbe6-97d422d5f97e\" (UID: \"80610e88-4157-4a37-bbe6-97d422d5f97e\") " Oct 07 14:04:33 crc kubenswrapper[4959]: I1007 14:04:33.149587 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpkbg\" (UniqueName: \"kubernetes.io/projected/63e77afb-d7b9-46a0-8f3c-e12d8157812d-kube-api-access-hpkbg\") pod \"63e77afb-d7b9-46a0-8f3c-e12d8157812d\" (UID: \"63e77afb-d7b9-46a0-8f3c-e12d8157812d\") " Oct 07 14:04:33 crc kubenswrapper[4959]: I1007 14:04:33.153964 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80610e88-4157-4a37-bbe6-97d422d5f97e-kube-api-access-9q4kb" (OuterVolumeSpecName: "kube-api-access-9q4kb") pod "80610e88-4157-4a37-bbe6-97d422d5f97e" (UID: "80610e88-4157-4a37-bbe6-97d422d5f97e"). InnerVolumeSpecName "kube-api-access-9q4kb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:04:33 crc kubenswrapper[4959]: I1007 14:04:33.154908 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63e77afb-d7b9-46a0-8f3c-e12d8157812d-kube-api-access-hpkbg" (OuterVolumeSpecName: "kube-api-access-hpkbg") pod "63e77afb-d7b9-46a0-8f3c-e12d8157812d" (UID: "63e77afb-d7b9-46a0-8f3c-e12d8157812d"). InnerVolumeSpecName "kube-api-access-hpkbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:04:33 crc kubenswrapper[4959]: I1007 14:04:33.251296 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9q4kb\" (UniqueName: \"kubernetes.io/projected/80610e88-4157-4a37-bbe6-97d422d5f97e-kube-api-access-9q4kb\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:33 crc kubenswrapper[4959]: I1007 14:04:33.251351 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpkbg\" (UniqueName: \"kubernetes.io/projected/63e77afb-d7b9-46a0-8f3c-e12d8157812d-kube-api-access-hpkbg\") on node \"crc\" DevicePath \"\"" Oct 07 14:04:33 crc kubenswrapper[4959]: I1007 14:04:33.567961 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-1479-account-create-n9c4j" event={"ID":"63e77afb-d7b9-46a0-8f3c-e12d8157812d","Type":"ContainerDied","Data":"acdefb3260178dac638f860632af6e11a8698d29c85b46a8bae3f098f90cc003"} Oct 07 14:04:33 crc kubenswrapper[4959]: I1007 14:04:33.568005 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="acdefb3260178dac638f860632af6e11a8698d29c85b46a8bae3f098f90cc003" Oct 07 14:04:33 crc kubenswrapper[4959]: I1007 14:04:33.568034 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-1479-account-create-n9c4j" Oct 07 14:04:33 crc kubenswrapper[4959]: I1007 14:04:33.580667 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-79f7-account-create-7n7bb" Oct 07 14:04:33 crc kubenswrapper[4959]: I1007 14:04:33.580673 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-79f7-account-create-7n7bb" event={"ID":"80610e88-4157-4a37-bbe6-97d422d5f97e","Type":"ContainerDied","Data":"a5c73a999de63e8813ce9fd5cd0ee2995e6189dfa72b1fa575de0cf9f385fe65"} Oct 07 14:04:33 crc kubenswrapper[4959]: I1007 14:04:33.580714 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5c73a999de63e8813ce9fd5cd0ee2995e6189dfa72b1fa575de0cf9f385fe65" Oct 07 14:04:33 crc kubenswrapper[4959]: I1007 14:04:33.582369 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f0b3a16-13a7-42ab-82a6-f2171da788d3","Type":"ContainerStarted","Data":"1ae6f3fd34e806edd1ed915695cf6d8b70ecc999e759c93cf244d325c348cab6"} Oct 07 14:04:35 crc kubenswrapper[4959]: I1007 14:04:35.546267 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6djrf"] Oct 07 14:04:35 crc kubenswrapper[4959]: E1007 14:04:35.546945 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80610e88-4157-4a37-bbe6-97d422d5f97e" containerName="mariadb-account-create" Oct 07 14:04:35 crc kubenswrapper[4959]: I1007 14:04:35.546963 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="80610e88-4157-4a37-bbe6-97d422d5f97e" containerName="mariadb-account-create" Oct 07 14:04:35 crc kubenswrapper[4959]: E1007 14:04:35.546992 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63e77afb-d7b9-46a0-8f3c-e12d8157812d" containerName="mariadb-account-create" Oct 07 14:04:35 crc kubenswrapper[4959]: I1007 14:04:35.547003 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="63e77afb-d7b9-46a0-8f3c-e12d8157812d" containerName="mariadb-account-create" Oct 07 14:04:35 crc kubenswrapper[4959]: E1007 14:04:35.547046 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a588fbdc-fd65-49f6-ab59-4901c7ca690a" containerName="horizon-log" Oct 07 14:04:35 crc kubenswrapper[4959]: I1007 14:04:35.547058 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a588fbdc-fd65-49f6-ab59-4901c7ca690a" containerName="horizon-log" Oct 07 14:04:35 crc kubenswrapper[4959]: E1007 14:04:35.547078 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a588fbdc-fd65-49f6-ab59-4901c7ca690a" containerName="horizon" Oct 07 14:04:35 crc kubenswrapper[4959]: I1007 14:04:35.547088 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a588fbdc-fd65-49f6-ab59-4901c7ca690a" containerName="horizon" Oct 07 14:04:35 crc kubenswrapper[4959]: I1007 14:04:35.547305 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a588fbdc-fd65-49f6-ab59-4901c7ca690a" containerName="horizon" Oct 07 14:04:35 crc kubenswrapper[4959]: I1007 14:04:35.547330 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="63e77afb-d7b9-46a0-8f3c-e12d8157812d" containerName="mariadb-account-create" Oct 07 14:04:35 crc kubenswrapper[4959]: I1007 14:04:35.547347 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a588fbdc-fd65-49f6-ab59-4901c7ca690a" containerName="horizon-log" Oct 07 14:04:35 crc kubenswrapper[4959]: I1007 14:04:35.547367 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="80610e88-4157-4a37-bbe6-97d422d5f97e" containerName="mariadb-account-create" Oct 07 14:04:35 crc kubenswrapper[4959]: I1007 14:04:35.548038 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6djrf" Oct 07 14:04:35 crc kubenswrapper[4959]: I1007 14:04:35.551113 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 07 14:04:35 crc kubenswrapper[4959]: I1007 14:04:35.551520 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-gn7n2" Oct 07 14:04:35 crc kubenswrapper[4959]: I1007 14:04:35.551588 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 07 14:04:35 crc kubenswrapper[4959]: I1007 14:04:35.556718 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6djrf"] Oct 07 14:04:35 crc kubenswrapper[4959]: I1007 14:04:35.691723 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bfaeeb0-e041-452a-aa79-9a65f2b1def8-config-data\") pod \"nova-cell0-conductor-db-sync-6djrf\" (UID: \"7bfaeeb0-e041-452a-aa79-9a65f2b1def8\") " pod="openstack/nova-cell0-conductor-db-sync-6djrf" Oct 07 14:04:35 crc kubenswrapper[4959]: I1007 14:04:35.691992 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bfaeeb0-e041-452a-aa79-9a65f2b1def8-scripts\") pod \"nova-cell0-conductor-db-sync-6djrf\" (UID: \"7bfaeeb0-e041-452a-aa79-9a65f2b1def8\") " pod="openstack/nova-cell0-conductor-db-sync-6djrf" Oct 07 14:04:35 crc kubenswrapper[4959]: I1007 14:04:35.692177 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtmdt\" (UniqueName: \"kubernetes.io/projected/7bfaeeb0-e041-452a-aa79-9a65f2b1def8-kube-api-access-xtmdt\") pod \"nova-cell0-conductor-db-sync-6djrf\" (UID: \"7bfaeeb0-e041-452a-aa79-9a65f2b1def8\") " pod="openstack/nova-cell0-conductor-db-sync-6djrf" Oct 07 14:04:35 crc kubenswrapper[4959]: I1007 14:04:35.692321 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bfaeeb0-e041-452a-aa79-9a65f2b1def8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-6djrf\" (UID: \"7bfaeeb0-e041-452a-aa79-9a65f2b1def8\") " pod="openstack/nova-cell0-conductor-db-sync-6djrf" Oct 07 14:04:35 crc kubenswrapper[4959]: I1007 14:04:35.793977 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bfaeeb0-e041-452a-aa79-9a65f2b1def8-config-data\") pod \"nova-cell0-conductor-db-sync-6djrf\" (UID: \"7bfaeeb0-e041-452a-aa79-9a65f2b1def8\") " pod="openstack/nova-cell0-conductor-db-sync-6djrf" Oct 07 14:04:35 crc kubenswrapper[4959]: I1007 14:04:35.794033 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bfaeeb0-e041-452a-aa79-9a65f2b1def8-scripts\") pod \"nova-cell0-conductor-db-sync-6djrf\" (UID: \"7bfaeeb0-e041-452a-aa79-9a65f2b1def8\") " pod="openstack/nova-cell0-conductor-db-sync-6djrf" Oct 07 14:04:35 crc kubenswrapper[4959]: I1007 14:04:35.794053 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtmdt\" (UniqueName: \"kubernetes.io/projected/7bfaeeb0-e041-452a-aa79-9a65f2b1def8-kube-api-access-xtmdt\") pod \"nova-cell0-conductor-db-sync-6djrf\" (UID: \"7bfaeeb0-e041-452a-aa79-9a65f2b1def8\") " pod="openstack/nova-cell0-conductor-db-sync-6djrf" Oct 07 14:04:35 crc kubenswrapper[4959]: I1007 14:04:35.794137 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bfaeeb0-e041-452a-aa79-9a65f2b1def8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-6djrf\" (UID: \"7bfaeeb0-e041-452a-aa79-9a65f2b1def8\") " pod="openstack/nova-cell0-conductor-db-sync-6djrf" Oct 07 14:04:35 crc kubenswrapper[4959]: I1007 14:04:35.797930 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bfaeeb0-e041-452a-aa79-9a65f2b1def8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-6djrf\" (UID: \"7bfaeeb0-e041-452a-aa79-9a65f2b1def8\") " pod="openstack/nova-cell0-conductor-db-sync-6djrf" Oct 07 14:04:35 crc kubenswrapper[4959]: I1007 14:04:35.798589 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bfaeeb0-e041-452a-aa79-9a65f2b1def8-scripts\") pod \"nova-cell0-conductor-db-sync-6djrf\" (UID: \"7bfaeeb0-e041-452a-aa79-9a65f2b1def8\") " pod="openstack/nova-cell0-conductor-db-sync-6djrf" Oct 07 14:04:35 crc kubenswrapper[4959]: I1007 14:04:35.800448 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bfaeeb0-e041-452a-aa79-9a65f2b1def8-config-data\") pod \"nova-cell0-conductor-db-sync-6djrf\" (UID: \"7bfaeeb0-e041-452a-aa79-9a65f2b1def8\") " pod="openstack/nova-cell0-conductor-db-sync-6djrf" Oct 07 14:04:35 crc kubenswrapper[4959]: I1007 14:04:35.812896 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtmdt\" (UniqueName: \"kubernetes.io/projected/7bfaeeb0-e041-452a-aa79-9a65f2b1def8-kube-api-access-xtmdt\") pod \"nova-cell0-conductor-db-sync-6djrf\" (UID: \"7bfaeeb0-e041-452a-aa79-9a65f2b1def8\") " pod="openstack/nova-cell0-conductor-db-sync-6djrf" Oct 07 14:04:35 crc kubenswrapper[4959]: I1007 14:04:35.972021 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6djrf" Oct 07 14:04:36 crc kubenswrapper[4959]: I1007 14:04:36.390422 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6djrf"] Oct 07 14:04:36 crc kubenswrapper[4959]: W1007 14:04:36.394748 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7bfaeeb0_e041_452a_aa79_9a65f2b1def8.slice/crio-47a5cf60fe8efc29e3aacdc2658417df71e59bf59f385940c45d2965117ccfee WatchSource:0}: Error finding container 47a5cf60fe8efc29e3aacdc2658417df71e59bf59f385940c45d2965117ccfee: Status 404 returned error can't find the container with id 47a5cf60fe8efc29e3aacdc2658417df71e59bf59f385940c45d2965117ccfee Oct 07 14:04:36 crc kubenswrapper[4959]: I1007 14:04:36.622811 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f0b3a16-13a7-42ab-82a6-f2171da788d3","Type":"ContainerStarted","Data":"b3e87fa0746241cbe6d827469c862d51bd056499ab0427d1d302d89773453ab2"} Oct 07 14:04:36 crc kubenswrapper[4959]: I1007 14:04:36.626450 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6djrf" event={"ID":"7bfaeeb0-e041-452a-aa79-9a65f2b1def8","Type":"ContainerStarted","Data":"47a5cf60fe8efc29e3aacdc2658417df71e59bf59f385940c45d2965117ccfee"} Oct 07 14:04:38 crc kubenswrapper[4959]: I1007 14:04:38.642159 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f0b3a16-13a7-42ab-82a6-f2171da788d3","Type":"ContainerStarted","Data":"a4d7bb33ececf070a1e9ea13eeab188249cd684ca396519b3a338f2f55796baf"} Oct 07 14:04:38 crc kubenswrapper[4959]: I1007 14:04:38.642730 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 07 14:04:38 crc kubenswrapper[4959]: I1007 14:04:38.665220 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.6918417789999998 podStartE2EDuration="8.665198068s" podCreationTimestamp="2025-10-07 14:04:30 +0000 UTC" firstStartedPulling="2025-10-07 14:04:31.335374635 +0000 UTC m=+1153.418778960" lastFinishedPulling="2025-10-07 14:04:37.308730924 +0000 UTC m=+1159.392135249" observedRunningTime="2025-10-07 14:04:38.658206075 +0000 UTC m=+1160.741610420" watchObservedRunningTime="2025-10-07 14:04:38.665198068 +0000 UTC m=+1160.748602393" Oct 07 14:04:44 crc kubenswrapper[4959]: I1007 14:04:44.694618 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6djrf" event={"ID":"7bfaeeb0-e041-452a-aa79-9a65f2b1def8","Type":"ContainerStarted","Data":"0d012cf781bb879e672e2a7388909f33c2f8516a32238cb1e064ce47451241ef"} Oct 07 14:04:44 crc kubenswrapper[4959]: I1007 14:04:44.714786 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-6djrf" podStartSLOduration=2.513676769 podStartE2EDuration="9.714769484s" podCreationTimestamp="2025-10-07 14:04:35 +0000 UTC" firstStartedPulling="2025-10-07 14:04:36.397457734 +0000 UTC m=+1158.480862059" lastFinishedPulling="2025-10-07 14:04:43.598550449 +0000 UTC m=+1165.681954774" observedRunningTime="2025-10-07 14:04:44.712075717 +0000 UTC m=+1166.795480062" watchObservedRunningTime="2025-10-07 14:04:44.714769484 +0000 UTC m=+1166.798173799" Oct 07 14:04:50 crc kubenswrapper[4959]: I1007 14:04:50.631386 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:04:50 crc kubenswrapper[4959]: I1007 14:04:50.632056 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:04:50 crc kubenswrapper[4959]: I1007 14:04:50.632142 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 14:04:50 crc kubenswrapper[4959]: I1007 14:04:50.633084 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f783f953eb3f6009a38800382e284161bac6530f08d9be67139e2b69d6c3e99d"} pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 14:04:50 crc kubenswrapper[4959]: I1007 14:04:50.633191 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" containerID="cri-o://f783f953eb3f6009a38800382e284161bac6530f08d9be67139e2b69d6c3e99d" gracePeriod=600 Oct 07 14:04:51 crc kubenswrapper[4959]: I1007 14:04:51.770063 4959 generic.go:334] "Generic (PLEG): container finished" podID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerID="f783f953eb3f6009a38800382e284161bac6530f08d9be67139e2b69d6c3e99d" exitCode=0 Oct 07 14:04:51 crc kubenswrapper[4959]: I1007 14:04:51.770234 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerDied","Data":"f783f953eb3f6009a38800382e284161bac6530f08d9be67139e2b69d6c3e99d"} Oct 07 14:04:51 crc kubenswrapper[4959]: I1007 14:04:51.770600 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerStarted","Data":"18f3e5766b69d97101a7543166ad26f606c175effe330b99db89cae18c7ad6ba"} Oct 07 14:04:51 crc kubenswrapper[4959]: I1007 14:04:51.770622 4959 scope.go:117] "RemoveContainer" containerID="be8cac39c0f211821bf3cae10b6e2f58082165f269df748f76ccd66a0c6a1240" Oct 07 14:05:00 crc kubenswrapper[4959]: I1007 14:05:00.916250 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 07 14:05:03 crc kubenswrapper[4959]: I1007 14:05:03.414817 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 14:05:03 crc kubenswrapper[4959]: I1007 14:05:03.415388 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="8b37b4cb-8660-456d-8d54-8e9c7fc83c51" containerName="kube-state-metrics" containerID="cri-o://7691c6913d8e8a52f64be65d70d9957ab25a7302b12a66f302caef479bc76ff1" gracePeriod=30 Oct 07 14:05:03 crc kubenswrapper[4959]: I1007 14:05:03.634234 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="8b37b4cb-8660-456d-8d54-8e9c7fc83c51" containerName="kube-state-metrics" probeResult="failure" output="Get \"http://10.217.0.106:8081/readyz\": dial tcp 10.217.0.106:8081: connect: connection refused" Oct 07 14:05:03 crc kubenswrapper[4959]: I1007 14:05:03.882453 4959 generic.go:334] "Generic (PLEG): container finished" podID="8b37b4cb-8660-456d-8d54-8e9c7fc83c51" containerID="7691c6913d8e8a52f64be65d70d9957ab25a7302b12a66f302caef479bc76ff1" exitCode=2 Oct 07 14:05:03 crc kubenswrapper[4959]: I1007 14:05:03.882497 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8b37b4cb-8660-456d-8d54-8e9c7fc83c51","Type":"ContainerDied","Data":"7691c6913d8e8a52f64be65d70d9957ab25a7302b12a66f302caef479bc76ff1"} Oct 07 14:05:04 crc kubenswrapper[4959]: I1007 14:05:04.371555 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 07 14:05:04 crc kubenswrapper[4959]: I1007 14:05:04.519998 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jbht\" (UniqueName: \"kubernetes.io/projected/8b37b4cb-8660-456d-8d54-8e9c7fc83c51-kube-api-access-2jbht\") pod \"8b37b4cb-8660-456d-8d54-8e9c7fc83c51\" (UID: \"8b37b4cb-8660-456d-8d54-8e9c7fc83c51\") " Oct 07 14:05:04 crc kubenswrapper[4959]: I1007 14:05:04.529325 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b37b4cb-8660-456d-8d54-8e9c7fc83c51-kube-api-access-2jbht" (OuterVolumeSpecName: "kube-api-access-2jbht") pod "8b37b4cb-8660-456d-8d54-8e9c7fc83c51" (UID: "8b37b4cb-8660-456d-8d54-8e9c7fc83c51"). InnerVolumeSpecName "kube-api-access-2jbht". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:05:04 crc kubenswrapper[4959]: I1007 14:05:04.562770 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:05:04 crc kubenswrapper[4959]: I1007 14:05:04.563362 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2f0b3a16-13a7-42ab-82a6-f2171da788d3" containerName="sg-core" containerID="cri-o://b3e87fa0746241cbe6d827469c862d51bd056499ab0427d1d302d89773453ab2" gracePeriod=30 Oct 07 14:05:04 crc kubenswrapper[4959]: I1007 14:05:04.563473 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2f0b3a16-13a7-42ab-82a6-f2171da788d3" containerName="proxy-httpd" containerID="cri-o://a4d7bb33ececf070a1e9ea13eeab188249cd684ca396519b3a338f2f55796baf" gracePeriod=30 Oct 07 14:05:04 crc kubenswrapper[4959]: I1007 14:05:04.563568 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2f0b3a16-13a7-42ab-82a6-f2171da788d3" containerName="ceilometer-notification-agent" containerID="cri-o://1ae6f3fd34e806edd1ed915695cf6d8b70ecc999e759c93cf244d325c348cab6" gracePeriod=30 Oct 07 14:05:04 crc kubenswrapper[4959]: I1007 14:05:04.563050 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2f0b3a16-13a7-42ab-82a6-f2171da788d3" containerName="ceilometer-central-agent" containerID="cri-o://5a3b72ba50e34ac0f36f43c40f436380ec0feedb7647c4127fdbaa9f05031e00" gracePeriod=30 Oct 07 14:05:04 crc kubenswrapper[4959]: I1007 14:05:04.622018 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jbht\" (UniqueName: \"kubernetes.io/projected/8b37b4cb-8660-456d-8d54-8e9c7fc83c51-kube-api-access-2jbht\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:04 crc kubenswrapper[4959]: I1007 14:05:04.897136 4959 generic.go:334] "Generic (PLEG): container finished" podID="2f0b3a16-13a7-42ab-82a6-f2171da788d3" containerID="a4d7bb33ececf070a1e9ea13eeab188249cd684ca396519b3a338f2f55796baf" exitCode=0 Oct 07 14:05:04 crc kubenswrapper[4959]: I1007 14:05:04.897186 4959 generic.go:334] "Generic (PLEG): container finished" podID="2f0b3a16-13a7-42ab-82a6-f2171da788d3" containerID="b3e87fa0746241cbe6d827469c862d51bd056499ab0427d1d302d89773453ab2" exitCode=2 Oct 07 14:05:04 crc kubenswrapper[4959]: I1007 14:05:04.897184 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f0b3a16-13a7-42ab-82a6-f2171da788d3","Type":"ContainerDied","Data":"a4d7bb33ececf070a1e9ea13eeab188249cd684ca396519b3a338f2f55796baf"} Oct 07 14:05:04 crc kubenswrapper[4959]: I1007 14:05:04.897242 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f0b3a16-13a7-42ab-82a6-f2171da788d3","Type":"ContainerDied","Data":"b3e87fa0746241cbe6d827469c862d51bd056499ab0427d1d302d89773453ab2"} Oct 07 14:05:04 crc kubenswrapper[4959]: I1007 14:05:04.899551 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8b37b4cb-8660-456d-8d54-8e9c7fc83c51","Type":"ContainerDied","Data":"70e8864f749c83dccd7480a3686e850af2d76b658c16cceb91cd1dfcd54d94e7"} Oct 07 14:05:04 crc kubenswrapper[4959]: I1007 14:05:04.899611 4959 scope.go:117] "RemoveContainer" containerID="7691c6913d8e8a52f64be65d70d9957ab25a7302b12a66f302caef479bc76ff1" Oct 07 14:05:04 crc kubenswrapper[4959]: I1007 14:05:04.899827 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 07 14:05:04 crc kubenswrapper[4959]: I1007 14:05:04.932150 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 14:05:04 crc kubenswrapper[4959]: I1007 14:05:04.941616 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 14:05:04 crc kubenswrapper[4959]: I1007 14:05:04.952918 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 14:05:04 crc kubenswrapper[4959]: E1007 14:05:04.953721 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b37b4cb-8660-456d-8d54-8e9c7fc83c51" containerName="kube-state-metrics" Oct 07 14:05:04 crc kubenswrapper[4959]: I1007 14:05:04.953741 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b37b4cb-8660-456d-8d54-8e9c7fc83c51" containerName="kube-state-metrics" Oct 07 14:05:04 crc kubenswrapper[4959]: I1007 14:05:04.953918 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b37b4cb-8660-456d-8d54-8e9c7fc83c51" containerName="kube-state-metrics" Oct 07 14:05:04 crc kubenswrapper[4959]: I1007 14:05:04.954505 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 07 14:05:04 crc kubenswrapper[4959]: I1007 14:05:04.960462 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 07 14:05:04 crc kubenswrapper[4959]: I1007 14:05:04.960474 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 07 14:05:04 crc kubenswrapper[4959]: I1007 14:05:04.968317 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.130812 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/8b8dd772-3ff0-474c-8571-ddf2bed03a82-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"8b8dd772-3ff0-474c-8571-ddf2bed03a82\") " pod="openstack/kube-state-metrics-0" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.130936 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7css\" (UniqueName: \"kubernetes.io/projected/8b8dd772-3ff0-474c-8571-ddf2bed03a82-kube-api-access-p7css\") pod \"kube-state-metrics-0\" (UID: \"8b8dd772-3ff0-474c-8571-ddf2bed03a82\") " pod="openstack/kube-state-metrics-0" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.130964 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b8dd772-3ff0-474c-8571-ddf2bed03a82-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"8b8dd772-3ff0-474c-8571-ddf2bed03a82\") " pod="openstack/kube-state-metrics-0" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.130983 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b8dd772-3ff0-474c-8571-ddf2bed03a82-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"8b8dd772-3ff0-474c-8571-ddf2bed03a82\") " pod="openstack/kube-state-metrics-0" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.232702 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/8b8dd772-3ff0-474c-8571-ddf2bed03a82-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"8b8dd772-3ff0-474c-8571-ddf2bed03a82\") " pod="openstack/kube-state-metrics-0" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.232820 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7css\" (UniqueName: \"kubernetes.io/projected/8b8dd772-3ff0-474c-8571-ddf2bed03a82-kube-api-access-p7css\") pod \"kube-state-metrics-0\" (UID: \"8b8dd772-3ff0-474c-8571-ddf2bed03a82\") " pod="openstack/kube-state-metrics-0" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.232849 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b8dd772-3ff0-474c-8571-ddf2bed03a82-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"8b8dd772-3ff0-474c-8571-ddf2bed03a82\") " pod="openstack/kube-state-metrics-0" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.232865 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b8dd772-3ff0-474c-8571-ddf2bed03a82-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"8b8dd772-3ff0-474c-8571-ddf2bed03a82\") " pod="openstack/kube-state-metrics-0" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.237276 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/8b8dd772-3ff0-474c-8571-ddf2bed03a82-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"8b8dd772-3ff0-474c-8571-ddf2bed03a82\") " pod="openstack/kube-state-metrics-0" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.237507 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b8dd772-3ff0-474c-8571-ddf2bed03a82-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"8b8dd772-3ff0-474c-8571-ddf2bed03a82\") " pod="openstack/kube-state-metrics-0" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.247222 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b8dd772-3ff0-474c-8571-ddf2bed03a82-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"8b8dd772-3ff0-474c-8571-ddf2bed03a82\") " pod="openstack/kube-state-metrics-0" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.258166 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7css\" (UniqueName: \"kubernetes.io/projected/8b8dd772-3ff0-474c-8571-ddf2bed03a82-kube-api-access-p7css\") pod \"kube-state-metrics-0\" (UID: \"8b8dd772-3ff0-474c-8571-ddf2bed03a82\") " pod="openstack/kube-state-metrics-0" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.331415 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.470365 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.540813 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f0b3a16-13a7-42ab-82a6-f2171da788d3-log-httpd\") pod \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\" (UID: \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\") " Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.540961 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f0b3a16-13a7-42ab-82a6-f2171da788d3-combined-ca-bundle\") pod \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\" (UID: \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\") " Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.541088 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f954n\" (UniqueName: \"kubernetes.io/projected/2f0b3a16-13a7-42ab-82a6-f2171da788d3-kube-api-access-f954n\") pod \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\" (UID: \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\") " Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.541143 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f0b3a16-13a7-42ab-82a6-f2171da788d3-scripts\") pod \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\" (UID: \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\") " Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.541176 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f0b3a16-13a7-42ab-82a6-f2171da788d3-run-httpd\") pod \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\" (UID: \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\") " Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.541532 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f0b3a16-13a7-42ab-82a6-f2171da788d3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2f0b3a16-13a7-42ab-82a6-f2171da788d3" (UID: "2f0b3a16-13a7-42ab-82a6-f2171da788d3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.541586 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f0b3a16-13a7-42ab-82a6-f2171da788d3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2f0b3a16-13a7-42ab-82a6-f2171da788d3" (UID: "2f0b3a16-13a7-42ab-82a6-f2171da788d3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.541896 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f0b3a16-13a7-42ab-82a6-f2171da788d3-config-data\") pod \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\" (UID: \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\") " Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.541977 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f0b3a16-13a7-42ab-82a6-f2171da788d3-sg-core-conf-yaml\") pod \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\" (UID: \"2f0b3a16-13a7-42ab-82a6-f2171da788d3\") " Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.542392 4959 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f0b3a16-13a7-42ab-82a6-f2171da788d3-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.542410 4959 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f0b3a16-13a7-42ab-82a6-f2171da788d3-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.546251 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f0b3a16-13a7-42ab-82a6-f2171da788d3-kube-api-access-f954n" (OuterVolumeSpecName: "kube-api-access-f954n") pod "2f0b3a16-13a7-42ab-82a6-f2171da788d3" (UID: "2f0b3a16-13a7-42ab-82a6-f2171da788d3"). InnerVolumeSpecName "kube-api-access-f954n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.546917 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f0b3a16-13a7-42ab-82a6-f2171da788d3-scripts" (OuterVolumeSpecName: "scripts") pod "2f0b3a16-13a7-42ab-82a6-f2171da788d3" (UID: "2f0b3a16-13a7-42ab-82a6-f2171da788d3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.597511 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f0b3a16-13a7-42ab-82a6-f2171da788d3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2f0b3a16-13a7-42ab-82a6-f2171da788d3" (UID: "2f0b3a16-13a7-42ab-82a6-f2171da788d3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.612313 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f0b3a16-13a7-42ab-82a6-f2171da788d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2f0b3a16-13a7-42ab-82a6-f2171da788d3" (UID: "2f0b3a16-13a7-42ab-82a6-f2171da788d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.643644 4959 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f0b3a16-13a7-42ab-82a6-f2171da788d3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.643681 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f0b3a16-13a7-42ab-82a6-f2171da788d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.643748 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f954n\" (UniqueName: \"kubernetes.io/projected/2f0b3a16-13a7-42ab-82a6-f2171da788d3-kube-api-access-f954n\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.643773 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f0b3a16-13a7-42ab-82a6-f2171da788d3-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.658647 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f0b3a16-13a7-42ab-82a6-f2171da788d3-config-data" (OuterVolumeSpecName: "config-data") pod "2f0b3a16-13a7-42ab-82a6-f2171da788d3" (UID: "2f0b3a16-13a7-42ab-82a6-f2171da788d3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.745127 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f0b3a16-13a7-42ab-82a6-f2171da788d3-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.801797 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.910958 4959 generic.go:334] "Generic (PLEG): container finished" podID="2f0b3a16-13a7-42ab-82a6-f2171da788d3" containerID="1ae6f3fd34e806edd1ed915695cf6d8b70ecc999e759c93cf244d325c348cab6" exitCode=0 Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.910997 4959 generic.go:334] "Generic (PLEG): container finished" podID="2f0b3a16-13a7-42ab-82a6-f2171da788d3" containerID="5a3b72ba50e34ac0f36f43c40f436380ec0feedb7647c4127fdbaa9f05031e00" exitCode=0 Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.911049 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f0b3a16-13a7-42ab-82a6-f2171da788d3","Type":"ContainerDied","Data":"1ae6f3fd34e806edd1ed915695cf6d8b70ecc999e759c93cf244d325c348cab6"} Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.911076 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f0b3a16-13a7-42ab-82a6-f2171da788d3","Type":"ContainerDied","Data":"5a3b72ba50e34ac0f36f43c40f436380ec0feedb7647c4127fdbaa9f05031e00"} Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.911086 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f0b3a16-13a7-42ab-82a6-f2171da788d3","Type":"ContainerDied","Data":"1db2f62deebe35c37375ed2df0ef70dd7086d3c1c867ed3707e39b2a2e8d36a6"} Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.911115 4959 scope.go:117] "RemoveContainer" containerID="a4d7bb33ececf070a1e9ea13eeab188249cd684ca396519b3a338f2f55796baf" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.911574 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.913225 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8b8dd772-3ff0-474c-8571-ddf2bed03a82","Type":"ContainerStarted","Data":"da30f61a0e08c1fcf7b604f7fb2f5af0cfd2f78c2247e74295d1ac1c7af25995"} Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.915718 4959 generic.go:334] "Generic (PLEG): container finished" podID="7bfaeeb0-e041-452a-aa79-9a65f2b1def8" containerID="0d012cf781bb879e672e2a7388909f33c2f8516a32238cb1e064ce47451241ef" exitCode=0 Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.915766 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6djrf" event={"ID":"7bfaeeb0-e041-452a-aa79-9a65f2b1def8","Type":"ContainerDied","Data":"0d012cf781bb879e672e2a7388909f33c2f8516a32238cb1e064ce47451241ef"} Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.940692 4959 scope.go:117] "RemoveContainer" containerID="b3e87fa0746241cbe6d827469c862d51bd056499ab0427d1d302d89773453ab2" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.956778 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.965124 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.969682 4959 scope.go:117] "RemoveContainer" containerID="1ae6f3fd34e806edd1ed915695cf6d8b70ecc999e759c93cf244d325c348cab6" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.980468 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:05:05 crc kubenswrapper[4959]: E1007 14:05:05.981462 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f0b3a16-13a7-42ab-82a6-f2171da788d3" containerName="ceilometer-central-agent" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.981485 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f0b3a16-13a7-42ab-82a6-f2171da788d3" containerName="ceilometer-central-agent" Oct 07 14:05:05 crc kubenswrapper[4959]: E1007 14:05:05.981514 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f0b3a16-13a7-42ab-82a6-f2171da788d3" containerName="proxy-httpd" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.981523 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f0b3a16-13a7-42ab-82a6-f2171da788d3" containerName="proxy-httpd" Oct 07 14:05:05 crc kubenswrapper[4959]: E1007 14:05:05.981537 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f0b3a16-13a7-42ab-82a6-f2171da788d3" containerName="sg-core" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.981547 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f0b3a16-13a7-42ab-82a6-f2171da788d3" containerName="sg-core" Oct 07 14:05:05 crc kubenswrapper[4959]: E1007 14:05:05.981569 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f0b3a16-13a7-42ab-82a6-f2171da788d3" containerName="ceilometer-notification-agent" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.981577 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f0b3a16-13a7-42ab-82a6-f2171da788d3" containerName="ceilometer-notification-agent" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.981768 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f0b3a16-13a7-42ab-82a6-f2171da788d3" containerName="sg-core" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.981792 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f0b3a16-13a7-42ab-82a6-f2171da788d3" containerName="proxy-httpd" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.981808 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f0b3a16-13a7-42ab-82a6-f2171da788d3" containerName="ceilometer-notification-agent" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.981828 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f0b3a16-13a7-42ab-82a6-f2171da788d3" containerName="ceilometer-central-agent" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.984462 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.988402 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.990219 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 07 14:05:05 crc kubenswrapper[4959]: I1007 14:05:05.990447 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.001167 4959 scope.go:117] "RemoveContainer" containerID="5a3b72ba50e34ac0f36f43c40f436380ec0feedb7647c4127fdbaa9f05031e00" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.085826 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80b77f5f-b572-4777-9912-800cc7886ddc-log-httpd\") pod \"ceilometer-0\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " pod="openstack/ceilometer-0" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.085866 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-scripts\") pod \"ceilometer-0\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " pod="openstack/ceilometer-0" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.085885 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80b77f5f-b572-4777-9912-800cc7886ddc-run-httpd\") pod \"ceilometer-0\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " pod="openstack/ceilometer-0" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.085902 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " pod="openstack/ceilometer-0" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.085943 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " pod="openstack/ceilometer-0" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.086046 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwqsr\" (UniqueName: \"kubernetes.io/projected/80b77f5f-b572-4777-9912-800cc7886ddc-kube-api-access-bwqsr\") pod \"ceilometer-0\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " pod="openstack/ceilometer-0" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.086070 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-config-data\") pod \"ceilometer-0\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " pod="openstack/ceilometer-0" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.086113 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " pod="openstack/ceilometer-0" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.088234 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.109266 4959 scope.go:117] "RemoveContainer" containerID="a4d7bb33ececf070a1e9ea13eeab188249cd684ca396519b3a338f2f55796baf" Oct 07 14:05:06 crc kubenswrapper[4959]: E1007 14:05:06.109608 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4d7bb33ececf070a1e9ea13eeab188249cd684ca396519b3a338f2f55796baf\": container with ID starting with a4d7bb33ececf070a1e9ea13eeab188249cd684ca396519b3a338f2f55796baf not found: ID does not exist" containerID="a4d7bb33ececf070a1e9ea13eeab188249cd684ca396519b3a338f2f55796baf" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.109650 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4d7bb33ececf070a1e9ea13eeab188249cd684ca396519b3a338f2f55796baf"} err="failed to get container status \"a4d7bb33ececf070a1e9ea13eeab188249cd684ca396519b3a338f2f55796baf\": rpc error: code = NotFound desc = could not find container \"a4d7bb33ececf070a1e9ea13eeab188249cd684ca396519b3a338f2f55796baf\": container with ID starting with a4d7bb33ececf070a1e9ea13eeab188249cd684ca396519b3a338f2f55796baf not found: ID does not exist" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.109678 4959 scope.go:117] "RemoveContainer" containerID="b3e87fa0746241cbe6d827469c862d51bd056499ab0427d1d302d89773453ab2" Oct 07 14:05:06 crc kubenswrapper[4959]: E1007 14:05:06.109899 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3e87fa0746241cbe6d827469c862d51bd056499ab0427d1d302d89773453ab2\": container with ID starting with b3e87fa0746241cbe6d827469c862d51bd056499ab0427d1d302d89773453ab2 not found: ID does not exist" containerID="b3e87fa0746241cbe6d827469c862d51bd056499ab0427d1d302d89773453ab2" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.109926 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3e87fa0746241cbe6d827469c862d51bd056499ab0427d1d302d89773453ab2"} err="failed to get container status \"b3e87fa0746241cbe6d827469c862d51bd056499ab0427d1d302d89773453ab2\": rpc error: code = NotFound desc = could not find container \"b3e87fa0746241cbe6d827469c862d51bd056499ab0427d1d302d89773453ab2\": container with ID starting with b3e87fa0746241cbe6d827469c862d51bd056499ab0427d1d302d89773453ab2 not found: ID does not exist" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.109948 4959 scope.go:117] "RemoveContainer" containerID="1ae6f3fd34e806edd1ed915695cf6d8b70ecc999e759c93cf244d325c348cab6" Oct 07 14:05:06 crc kubenswrapper[4959]: E1007 14:05:06.110190 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ae6f3fd34e806edd1ed915695cf6d8b70ecc999e759c93cf244d325c348cab6\": container with ID starting with 1ae6f3fd34e806edd1ed915695cf6d8b70ecc999e759c93cf244d325c348cab6 not found: ID does not exist" containerID="1ae6f3fd34e806edd1ed915695cf6d8b70ecc999e759c93cf244d325c348cab6" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.110209 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ae6f3fd34e806edd1ed915695cf6d8b70ecc999e759c93cf244d325c348cab6"} err="failed to get container status \"1ae6f3fd34e806edd1ed915695cf6d8b70ecc999e759c93cf244d325c348cab6\": rpc error: code = NotFound desc = could not find container \"1ae6f3fd34e806edd1ed915695cf6d8b70ecc999e759c93cf244d325c348cab6\": container with ID starting with 1ae6f3fd34e806edd1ed915695cf6d8b70ecc999e759c93cf244d325c348cab6 not found: ID does not exist" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.110221 4959 scope.go:117] "RemoveContainer" containerID="5a3b72ba50e34ac0f36f43c40f436380ec0feedb7647c4127fdbaa9f05031e00" Oct 07 14:05:06 crc kubenswrapper[4959]: E1007 14:05:06.110417 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a3b72ba50e34ac0f36f43c40f436380ec0feedb7647c4127fdbaa9f05031e00\": container with ID starting with 5a3b72ba50e34ac0f36f43c40f436380ec0feedb7647c4127fdbaa9f05031e00 not found: ID does not exist" containerID="5a3b72ba50e34ac0f36f43c40f436380ec0feedb7647c4127fdbaa9f05031e00" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.110434 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a3b72ba50e34ac0f36f43c40f436380ec0feedb7647c4127fdbaa9f05031e00"} err="failed to get container status \"5a3b72ba50e34ac0f36f43c40f436380ec0feedb7647c4127fdbaa9f05031e00\": rpc error: code = NotFound desc = could not find container \"5a3b72ba50e34ac0f36f43c40f436380ec0feedb7647c4127fdbaa9f05031e00\": container with ID starting with 5a3b72ba50e34ac0f36f43c40f436380ec0feedb7647c4127fdbaa9f05031e00 not found: ID does not exist" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.110446 4959 scope.go:117] "RemoveContainer" containerID="a4d7bb33ececf070a1e9ea13eeab188249cd684ca396519b3a338f2f55796baf" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.110640 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4d7bb33ececf070a1e9ea13eeab188249cd684ca396519b3a338f2f55796baf"} err="failed to get container status \"a4d7bb33ececf070a1e9ea13eeab188249cd684ca396519b3a338f2f55796baf\": rpc error: code = NotFound desc = could not find container \"a4d7bb33ececf070a1e9ea13eeab188249cd684ca396519b3a338f2f55796baf\": container with ID starting with a4d7bb33ececf070a1e9ea13eeab188249cd684ca396519b3a338f2f55796baf not found: ID does not exist" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.110658 4959 scope.go:117] "RemoveContainer" containerID="b3e87fa0746241cbe6d827469c862d51bd056499ab0427d1d302d89773453ab2" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.110817 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3e87fa0746241cbe6d827469c862d51bd056499ab0427d1d302d89773453ab2"} err="failed to get container status \"b3e87fa0746241cbe6d827469c862d51bd056499ab0427d1d302d89773453ab2\": rpc error: code = NotFound desc = could not find container \"b3e87fa0746241cbe6d827469c862d51bd056499ab0427d1d302d89773453ab2\": container with ID starting with b3e87fa0746241cbe6d827469c862d51bd056499ab0427d1d302d89773453ab2 not found: ID does not exist" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.110835 4959 scope.go:117] "RemoveContainer" containerID="1ae6f3fd34e806edd1ed915695cf6d8b70ecc999e759c93cf244d325c348cab6" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.110978 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ae6f3fd34e806edd1ed915695cf6d8b70ecc999e759c93cf244d325c348cab6"} err="failed to get container status \"1ae6f3fd34e806edd1ed915695cf6d8b70ecc999e759c93cf244d325c348cab6\": rpc error: code = NotFound desc = could not find container \"1ae6f3fd34e806edd1ed915695cf6d8b70ecc999e759c93cf244d325c348cab6\": container with ID starting with 1ae6f3fd34e806edd1ed915695cf6d8b70ecc999e759c93cf244d325c348cab6 not found: ID does not exist" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.110999 4959 scope.go:117] "RemoveContainer" containerID="5a3b72ba50e34ac0f36f43c40f436380ec0feedb7647c4127fdbaa9f05031e00" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.111164 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a3b72ba50e34ac0f36f43c40f436380ec0feedb7647c4127fdbaa9f05031e00"} err="failed to get container status \"5a3b72ba50e34ac0f36f43c40f436380ec0feedb7647c4127fdbaa9f05031e00\": rpc error: code = NotFound desc = could not find container \"5a3b72ba50e34ac0f36f43c40f436380ec0feedb7647c4127fdbaa9f05031e00\": container with ID starting with 5a3b72ba50e34ac0f36f43c40f436380ec0feedb7647c4127fdbaa9f05031e00 not found: ID does not exist" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.187306 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " pod="openstack/ceilometer-0" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.187361 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80b77f5f-b572-4777-9912-800cc7886ddc-log-httpd\") pod \"ceilometer-0\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " pod="openstack/ceilometer-0" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.187383 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-scripts\") pod \"ceilometer-0\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " pod="openstack/ceilometer-0" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.187404 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80b77f5f-b572-4777-9912-800cc7886ddc-run-httpd\") pod \"ceilometer-0\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " pod="openstack/ceilometer-0" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.187424 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " pod="openstack/ceilometer-0" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.187458 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " pod="openstack/ceilometer-0" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.187517 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwqsr\" (UniqueName: \"kubernetes.io/projected/80b77f5f-b572-4777-9912-800cc7886ddc-kube-api-access-bwqsr\") pod \"ceilometer-0\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " pod="openstack/ceilometer-0" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.187541 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-config-data\") pod \"ceilometer-0\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " pod="openstack/ceilometer-0" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.187915 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80b77f5f-b572-4777-9912-800cc7886ddc-run-httpd\") pod \"ceilometer-0\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " pod="openstack/ceilometer-0" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.188863 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80b77f5f-b572-4777-9912-800cc7886ddc-log-httpd\") pod \"ceilometer-0\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " pod="openstack/ceilometer-0" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.191246 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " pod="openstack/ceilometer-0" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.191383 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-config-data\") pod \"ceilometer-0\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " pod="openstack/ceilometer-0" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.192231 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " pod="openstack/ceilometer-0" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.192465 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " pod="openstack/ceilometer-0" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.192670 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-scripts\") pod \"ceilometer-0\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " pod="openstack/ceilometer-0" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.210504 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwqsr\" (UniqueName: \"kubernetes.io/projected/80b77f5f-b572-4777-9912-800cc7886ddc-kube-api-access-bwqsr\") pod \"ceilometer-0\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " pod="openstack/ceilometer-0" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.400706 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.663708 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f0b3a16-13a7-42ab-82a6-f2171da788d3" path="/var/lib/kubelet/pods/2f0b3a16-13a7-42ab-82a6-f2171da788d3/volumes" Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.664721 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b37b4cb-8660-456d-8d54-8e9c7fc83c51" path="/var/lib/kubelet/pods/8b37b4cb-8660-456d-8d54-8e9c7fc83c51/volumes" Oct 07 14:05:06 crc kubenswrapper[4959]: W1007 14:05:06.853185 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod80b77f5f_b572_4777_9912_800cc7886ddc.slice/crio-b99fe4719591640961f1eaca17de15e2e1481896e55e5b34efd6a6c76158b201 WatchSource:0}: Error finding container b99fe4719591640961f1eaca17de15e2e1481896e55e5b34efd6a6c76158b201: Status 404 returned error can't find the container with id b99fe4719591640961f1eaca17de15e2e1481896e55e5b34efd6a6c76158b201 Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.861229 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:05:06 crc kubenswrapper[4959]: I1007 14:05:06.924884 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"80b77f5f-b572-4777-9912-800cc7886ddc","Type":"ContainerStarted","Data":"b99fe4719591640961f1eaca17de15e2e1481896e55e5b34efd6a6c76158b201"} Oct 07 14:05:07 crc kubenswrapper[4959]: I1007 14:05:07.349254 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6djrf" Oct 07 14:05:07 crc kubenswrapper[4959]: I1007 14:05:07.574120 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtmdt\" (UniqueName: \"kubernetes.io/projected/7bfaeeb0-e041-452a-aa79-9a65f2b1def8-kube-api-access-xtmdt\") pod \"7bfaeeb0-e041-452a-aa79-9a65f2b1def8\" (UID: \"7bfaeeb0-e041-452a-aa79-9a65f2b1def8\") " Oct 07 14:05:07 crc kubenswrapper[4959]: I1007 14:05:07.574222 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bfaeeb0-e041-452a-aa79-9a65f2b1def8-scripts\") pod \"7bfaeeb0-e041-452a-aa79-9a65f2b1def8\" (UID: \"7bfaeeb0-e041-452a-aa79-9a65f2b1def8\") " Oct 07 14:05:07 crc kubenswrapper[4959]: I1007 14:05:07.574300 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bfaeeb0-e041-452a-aa79-9a65f2b1def8-config-data\") pod \"7bfaeeb0-e041-452a-aa79-9a65f2b1def8\" (UID: \"7bfaeeb0-e041-452a-aa79-9a65f2b1def8\") " Oct 07 14:05:07 crc kubenswrapper[4959]: I1007 14:05:07.574594 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bfaeeb0-e041-452a-aa79-9a65f2b1def8-combined-ca-bundle\") pod \"7bfaeeb0-e041-452a-aa79-9a65f2b1def8\" (UID: \"7bfaeeb0-e041-452a-aa79-9a65f2b1def8\") " Oct 07 14:05:07 crc kubenswrapper[4959]: I1007 14:05:07.578663 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bfaeeb0-e041-452a-aa79-9a65f2b1def8-kube-api-access-xtmdt" (OuterVolumeSpecName: "kube-api-access-xtmdt") pod "7bfaeeb0-e041-452a-aa79-9a65f2b1def8" (UID: "7bfaeeb0-e041-452a-aa79-9a65f2b1def8"). InnerVolumeSpecName "kube-api-access-xtmdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:05:07 crc kubenswrapper[4959]: I1007 14:05:07.579232 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bfaeeb0-e041-452a-aa79-9a65f2b1def8-scripts" (OuterVolumeSpecName: "scripts") pod "7bfaeeb0-e041-452a-aa79-9a65f2b1def8" (UID: "7bfaeeb0-e041-452a-aa79-9a65f2b1def8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:05:07 crc kubenswrapper[4959]: I1007 14:05:07.602835 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bfaeeb0-e041-452a-aa79-9a65f2b1def8-config-data" (OuterVolumeSpecName: "config-data") pod "7bfaeeb0-e041-452a-aa79-9a65f2b1def8" (UID: "7bfaeeb0-e041-452a-aa79-9a65f2b1def8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:05:07 crc kubenswrapper[4959]: I1007 14:05:07.611319 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bfaeeb0-e041-452a-aa79-9a65f2b1def8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7bfaeeb0-e041-452a-aa79-9a65f2b1def8" (UID: "7bfaeeb0-e041-452a-aa79-9a65f2b1def8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:05:07 crc kubenswrapper[4959]: I1007 14:05:07.676919 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bfaeeb0-e041-452a-aa79-9a65f2b1def8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:07 crc kubenswrapper[4959]: I1007 14:05:07.676963 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtmdt\" (UniqueName: \"kubernetes.io/projected/7bfaeeb0-e041-452a-aa79-9a65f2b1def8-kube-api-access-xtmdt\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:07 crc kubenswrapper[4959]: I1007 14:05:07.676980 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bfaeeb0-e041-452a-aa79-9a65f2b1def8-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:07 crc kubenswrapper[4959]: I1007 14:05:07.676991 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bfaeeb0-e041-452a-aa79-9a65f2b1def8-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:07 crc kubenswrapper[4959]: I1007 14:05:07.935089 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8b8dd772-3ff0-474c-8571-ddf2bed03a82","Type":"ContainerStarted","Data":"f43252a1bd43db6a4333225562664d5ab8a8a420ffb886a1e03d55ffb9f7ae37"} Oct 07 14:05:07 crc kubenswrapper[4959]: I1007 14:05:07.935689 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 07 14:05:07 crc kubenswrapper[4959]: I1007 14:05:07.936798 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6djrf" event={"ID":"7bfaeeb0-e041-452a-aa79-9a65f2b1def8","Type":"ContainerDied","Data":"47a5cf60fe8efc29e3aacdc2658417df71e59bf59f385940c45d2965117ccfee"} Oct 07 14:05:07 crc kubenswrapper[4959]: I1007 14:05:07.936838 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47a5cf60fe8efc29e3aacdc2658417df71e59bf59f385940c45d2965117ccfee" Oct 07 14:05:07 crc kubenswrapper[4959]: I1007 14:05:07.937092 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6djrf" Oct 07 14:05:07 crc kubenswrapper[4959]: I1007 14:05:07.964483 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.539293028 podStartE2EDuration="3.964463951s" podCreationTimestamp="2025-10-07 14:05:04 +0000 UTC" firstStartedPulling="2025-10-07 14:05:05.81108275 +0000 UTC m=+1187.894487075" lastFinishedPulling="2025-10-07 14:05:07.236253673 +0000 UTC m=+1189.319657998" observedRunningTime="2025-10-07 14:05:07.959679441 +0000 UTC m=+1190.043083786" watchObservedRunningTime="2025-10-07 14:05:07.964463951 +0000 UTC m=+1190.047868296" Oct 07 14:05:08 crc kubenswrapper[4959]: I1007 14:05:08.041143 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 07 14:05:08 crc kubenswrapper[4959]: E1007 14:05:08.041537 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bfaeeb0-e041-452a-aa79-9a65f2b1def8" containerName="nova-cell0-conductor-db-sync" Oct 07 14:05:08 crc kubenswrapper[4959]: I1007 14:05:08.041556 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bfaeeb0-e041-452a-aa79-9a65f2b1def8" containerName="nova-cell0-conductor-db-sync" Oct 07 14:05:08 crc kubenswrapper[4959]: I1007 14:05:08.041763 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bfaeeb0-e041-452a-aa79-9a65f2b1def8" containerName="nova-cell0-conductor-db-sync" Oct 07 14:05:08 crc kubenswrapper[4959]: I1007 14:05:08.042434 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 07 14:05:08 crc kubenswrapper[4959]: I1007 14:05:08.047313 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 07 14:05:08 crc kubenswrapper[4959]: I1007 14:05:08.047644 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-gn7n2" Oct 07 14:05:08 crc kubenswrapper[4959]: I1007 14:05:08.053809 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 07 14:05:08 crc kubenswrapper[4959]: I1007 14:05:08.084241 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74l4w\" (UniqueName: \"kubernetes.io/projected/ab63a619-32d4-4999-9056-cde8181e86ab-kube-api-access-74l4w\") pod \"nova-cell0-conductor-0\" (UID: \"ab63a619-32d4-4999-9056-cde8181e86ab\") " pod="openstack/nova-cell0-conductor-0" Oct 07 14:05:08 crc kubenswrapper[4959]: I1007 14:05:08.084505 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab63a619-32d4-4999-9056-cde8181e86ab-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"ab63a619-32d4-4999-9056-cde8181e86ab\") " pod="openstack/nova-cell0-conductor-0" Oct 07 14:05:08 crc kubenswrapper[4959]: I1007 14:05:08.084607 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab63a619-32d4-4999-9056-cde8181e86ab-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"ab63a619-32d4-4999-9056-cde8181e86ab\") " pod="openstack/nova-cell0-conductor-0" Oct 07 14:05:08 crc kubenswrapper[4959]: I1007 14:05:08.185940 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab63a619-32d4-4999-9056-cde8181e86ab-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"ab63a619-32d4-4999-9056-cde8181e86ab\") " pod="openstack/nova-cell0-conductor-0" Oct 07 14:05:08 crc kubenswrapper[4959]: I1007 14:05:08.185989 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab63a619-32d4-4999-9056-cde8181e86ab-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"ab63a619-32d4-4999-9056-cde8181e86ab\") " pod="openstack/nova-cell0-conductor-0" Oct 07 14:05:08 crc kubenswrapper[4959]: I1007 14:05:08.186146 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74l4w\" (UniqueName: \"kubernetes.io/projected/ab63a619-32d4-4999-9056-cde8181e86ab-kube-api-access-74l4w\") pod \"nova-cell0-conductor-0\" (UID: \"ab63a619-32d4-4999-9056-cde8181e86ab\") " pod="openstack/nova-cell0-conductor-0" Oct 07 14:05:08 crc kubenswrapper[4959]: I1007 14:05:08.191459 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab63a619-32d4-4999-9056-cde8181e86ab-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"ab63a619-32d4-4999-9056-cde8181e86ab\") " pod="openstack/nova-cell0-conductor-0" Oct 07 14:05:08 crc kubenswrapper[4959]: I1007 14:05:08.191473 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab63a619-32d4-4999-9056-cde8181e86ab-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"ab63a619-32d4-4999-9056-cde8181e86ab\") " pod="openstack/nova-cell0-conductor-0" Oct 07 14:05:08 crc kubenswrapper[4959]: I1007 14:05:08.205930 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74l4w\" (UniqueName: \"kubernetes.io/projected/ab63a619-32d4-4999-9056-cde8181e86ab-kube-api-access-74l4w\") pod \"nova-cell0-conductor-0\" (UID: \"ab63a619-32d4-4999-9056-cde8181e86ab\") " pod="openstack/nova-cell0-conductor-0" Oct 07 14:05:08 crc kubenswrapper[4959]: I1007 14:05:08.367270 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 07 14:05:08 crc kubenswrapper[4959]: I1007 14:05:08.840865 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 07 14:05:08 crc kubenswrapper[4959]: I1007 14:05:08.948812 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"80b77f5f-b572-4777-9912-800cc7886ddc","Type":"ContainerStarted","Data":"c1cacf38716e17dbbf7bd68462d448bdecbf3780bc758d010f17d82736b5c3ae"} Oct 07 14:05:08 crc kubenswrapper[4959]: I1007 14:05:08.950790 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"ab63a619-32d4-4999-9056-cde8181e86ab","Type":"ContainerStarted","Data":"4d82f65dd462ec9e4a81909dbad06131ac2dd6f12d3ddb3c6a391d3f262e658a"} Oct 07 14:05:09 crc kubenswrapper[4959]: I1007 14:05:09.961615 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"ab63a619-32d4-4999-9056-cde8181e86ab","Type":"ContainerStarted","Data":"07e8017c6feda3cdcd45bab209655f1eeeed19db89abc866a3087506456a8588"} Oct 07 14:05:09 crc kubenswrapper[4959]: I1007 14:05:09.962152 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 07 14:05:09 crc kubenswrapper[4959]: I1007 14:05:09.985180 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=1.9851612140000001 podStartE2EDuration="1.985161214s" podCreationTimestamp="2025-10-07 14:05:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:05:09.981534613 +0000 UTC m=+1192.064938938" watchObservedRunningTime="2025-10-07 14:05:09.985161214 +0000 UTC m=+1192.068565539" Oct 07 14:05:10 crc kubenswrapper[4959]: I1007 14:05:10.970606 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"80b77f5f-b572-4777-9912-800cc7886ddc","Type":"ContainerStarted","Data":"d083fcb14061bd5b6d3657744d575379b4a02f239d3e5703ba088b08b6136427"} Oct 07 14:05:11 crc kubenswrapper[4959]: I1007 14:05:11.979295 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"80b77f5f-b572-4777-9912-800cc7886ddc","Type":"ContainerStarted","Data":"394f73fb3149de229dcb4af397e77e4a465d315ea05e727bfa839d6025fcdb05"} Oct 07 14:05:15 crc kubenswrapper[4959]: I1007 14:05:15.021309 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"80b77f5f-b572-4777-9912-800cc7886ddc","Type":"ContainerStarted","Data":"a9eb1598f0d4f84a7bfc717906de92ea9e5a67b5f8a526d14bc6cb44f27d0100"} Oct 07 14:05:15 crc kubenswrapper[4959]: I1007 14:05:15.022483 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 07 14:05:15 crc kubenswrapper[4959]: I1007 14:05:15.342377 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 07 14:05:15 crc kubenswrapper[4959]: I1007 14:05:15.371965 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.110382379 podStartE2EDuration="10.371936024s" podCreationTimestamp="2025-10-07 14:05:05 +0000 UTC" firstStartedPulling="2025-10-07 14:05:06.955260777 +0000 UTC m=+1189.038665102" lastFinishedPulling="2025-10-07 14:05:14.216814422 +0000 UTC m=+1196.300218747" observedRunningTime="2025-10-07 14:05:15.047719935 +0000 UTC m=+1197.131124280" watchObservedRunningTime="2025-10-07 14:05:15.371936024 +0000 UTC m=+1197.455340389" Oct 07 14:05:18 crc kubenswrapper[4959]: I1007 14:05:18.393365 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 07 14:05:18 crc kubenswrapper[4959]: I1007 14:05:18.825205 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-qltzt"] Oct 07 14:05:18 crc kubenswrapper[4959]: I1007 14:05:18.826635 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qltzt" Oct 07 14:05:18 crc kubenswrapper[4959]: I1007 14:05:18.835986 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-qltzt"] Oct 07 14:05:18 crc kubenswrapper[4959]: I1007 14:05:18.837091 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 07 14:05:18 crc kubenswrapper[4959]: I1007 14:05:18.837301 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 07 14:05:18 crc kubenswrapper[4959]: I1007 14:05:18.993322 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvzzr\" (UniqueName: \"kubernetes.io/projected/1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f-kube-api-access-lvzzr\") pod \"nova-cell0-cell-mapping-qltzt\" (UID: \"1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f\") " pod="openstack/nova-cell0-cell-mapping-qltzt" Oct 07 14:05:18 crc kubenswrapper[4959]: I1007 14:05:18.993376 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qltzt\" (UID: \"1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f\") " pod="openstack/nova-cell0-cell-mapping-qltzt" Oct 07 14:05:18 crc kubenswrapper[4959]: I1007 14:05:18.993432 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f-config-data\") pod \"nova-cell0-cell-mapping-qltzt\" (UID: \"1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f\") " pod="openstack/nova-cell0-cell-mapping-qltzt" Oct 07 14:05:18 crc kubenswrapper[4959]: I1007 14:05:18.993582 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f-scripts\") pod \"nova-cell0-cell-mapping-qltzt\" (UID: \"1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f\") " pod="openstack/nova-cell0-cell-mapping-qltzt" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.015029 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.016684 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.023677 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.041145 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.076943 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.078464 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.085494 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.087346 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.094967 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xk2s\" (UniqueName: \"kubernetes.io/projected/6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4-kube-api-access-4xk2s\") pod \"nova-metadata-0\" (UID: \"6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4\") " pod="openstack/nova-metadata-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.095046 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvzzr\" (UniqueName: \"kubernetes.io/projected/1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f-kube-api-access-lvzzr\") pod \"nova-cell0-cell-mapping-qltzt\" (UID: \"1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f\") " pod="openstack/nova-cell0-cell-mapping-qltzt" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.095077 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qltzt\" (UID: \"1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f\") " pod="openstack/nova-cell0-cell-mapping-qltzt" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.095113 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fbef70d-f88f-4a9b-bbfd-cabbba6ab107-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1fbef70d-f88f-4a9b-bbfd-cabbba6ab107\") " pod="openstack/nova-api-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.095151 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4\") " pod="openstack/nova-metadata-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.095181 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4-config-data\") pod \"nova-metadata-0\" (UID: \"6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4\") " pod="openstack/nova-metadata-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.095201 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f-config-data\") pod \"nova-cell0-cell-mapping-qltzt\" (UID: \"1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f\") " pod="openstack/nova-cell0-cell-mapping-qltzt" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.095260 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1fbef70d-f88f-4a9b-bbfd-cabbba6ab107-logs\") pod \"nova-api-0\" (UID: \"1fbef70d-f88f-4a9b-bbfd-cabbba6ab107\") " pod="openstack/nova-api-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.095283 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9jvt\" (UniqueName: \"kubernetes.io/projected/1fbef70d-f88f-4a9b-bbfd-cabbba6ab107-kube-api-access-l9jvt\") pod \"nova-api-0\" (UID: \"1fbef70d-f88f-4a9b-bbfd-cabbba6ab107\") " pod="openstack/nova-api-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.095299 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fbef70d-f88f-4a9b-bbfd-cabbba6ab107-config-data\") pod \"nova-api-0\" (UID: \"1fbef70d-f88f-4a9b-bbfd-cabbba6ab107\") " pod="openstack/nova-api-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.095505 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4-logs\") pod \"nova-metadata-0\" (UID: \"6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4\") " pod="openstack/nova-metadata-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.095557 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f-scripts\") pod \"nova-cell0-cell-mapping-qltzt\" (UID: \"1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f\") " pod="openstack/nova-cell0-cell-mapping-qltzt" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.101816 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qltzt\" (UID: \"1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f\") " pod="openstack/nova-cell0-cell-mapping-qltzt" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.108549 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f-scripts\") pod \"nova-cell0-cell-mapping-qltzt\" (UID: \"1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f\") " pod="openstack/nova-cell0-cell-mapping-qltzt" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.118803 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f-config-data\") pod \"nova-cell0-cell-mapping-qltzt\" (UID: \"1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f\") " pod="openstack/nova-cell0-cell-mapping-qltzt" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.133632 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvzzr\" (UniqueName: \"kubernetes.io/projected/1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f-kube-api-access-lvzzr\") pod \"nova-cell0-cell-mapping-qltzt\" (UID: \"1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f\") " pod="openstack/nova-cell0-cell-mapping-qltzt" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.189915 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.191031 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.192811 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qltzt" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.196987 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fbef70d-f88f-4a9b-bbfd-cabbba6ab107-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1fbef70d-f88f-4a9b-bbfd-cabbba6ab107\") " pod="openstack/nova-api-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.197046 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4\") " pod="openstack/nova-metadata-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.197072 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4-config-data\") pod \"nova-metadata-0\" (UID: \"6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4\") " pod="openstack/nova-metadata-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.197138 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1fbef70d-f88f-4a9b-bbfd-cabbba6ab107-logs\") pod \"nova-api-0\" (UID: \"1fbef70d-f88f-4a9b-bbfd-cabbba6ab107\") " pod="openstack/nova-api-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.197162 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9jvt\" (UniqueName: \"kubernetes.io/projected/1fbef70d-f88f-4a9b-bbfd-cabbba6ab107-kube-api-access-l9jvt\") pod \"nova-api-0\" (UID: \"1fbef70d-f88f-4a9b-bbfd-cabbba6ab107\") " pod="openstack/nova-api-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.197176 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fbef70d-f88f-4a9b-bbfd-cabbba6ab107-config-data\") pod \"nova-api-0\" (UID: \"1fbef70d-f88f-4a9b-bbfd-cabbba6ab107\") " pod="openstack/nova-api-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.197203 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4-logs\") pod \"nova-metadata-0\" (UID: \"6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4\") " pod="openstack/nova-metadata-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.197226 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xk2s\" (UniqueName: \"kubernetes.io/projected/6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4-kube-api-access-4xk2s\") pod \"nova-metadata-0\" (UID: \"6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4\") " pod="openstack/nova-metadata-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.197755 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1fbef70d-f88f-4a9b-bbfd-cabbba6ab107-logs\") pod \"nova-api-0\" (UID: \"1fbef70d-f88f-4a9b-bbfd-cabbba6ab107\") " pod="openstack/nova-api-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.204885 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.205753 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4-logs\") pod \"nova-metadata-0\" (UID: \"6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4\") " pod="openstack/nova-metadata-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.206009 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4\") " pod="openstack/nova-metadata-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.206395 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4-config-data\") pod \"nova-metadata-0\" (UID: \"6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4\") " pod="openstack/nova-metadata-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.206526 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fbef70d-f88f-4a9b-bbfd-cabbba6ab107-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1fbef70d-f88f-4a9b-bbfd-cabbba6ab107\") " pod="openstack/nova-api-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.225160 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.236092 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fbef70d-f88f-4a9b-bbfd-cabbba6ab107-config-data\") pod \"nova-api-0\" (UID: \"1fbef70d-f88f-4a9b-bbfd-cabbba6ab107\") " pod="openstack/nova-api-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.247180 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.248479 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.251359 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.258778 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9jvt\" (UniqueName: \"kubernetes.io/projected/1fbef70d-f88f-4a9b-bbfd-cabbba6ab107-kube-api-access-l9jvt\") pod \"nova-api-0\" (UID: \"1fbef70d-f88f-4a9b-bbfd-cabbba6ab107\") " pod="openstack/nova-api-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.265936 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xk2s\" (UniqueName: \"kubernetes.io/projected/6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4-kube-api-access-4xk2s\") pod \"nova-metadata-0\" (UID: \"6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4\") " pod="openstack/nova-metadata-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.285441 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.298555 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd0e43f7-cac1-4f80-a489-9e9fb9c730b5-config-data\") pod \"nova-scheduler-0\" (UID: \"cd0e43f7-cac1-4f80-a489-9e9fb9c730b5\") " pod="openstack/nova-scheduler-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.298661 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pw8gj\" (UniqueName: \"kubernetes.io/projected/cd0e43f7-cac1-4f80-a489-9e9fb9c730b5-kube-api-access-pw8gj\") pod \"nova-scheduler-0\" (UID: \"cd0e43f7-cac1-4f80-a489-9e9fb9c730b5\") " pod="openstack/nova-scheduler-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.298724 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd0e43f7-cac1-4f80-a489-9e9fb9c730b5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cd0e43f7-cac1-4f80-a489-9e9fb9c730b5\") " pod="openstack/nova-scheduler-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.317160 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75fb48c489-xpjqw"] Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.318528 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75fb48c489-xpjqw" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.324934 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75fb48c489-xpjqw"] Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.341060 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.399779 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd0e43f7-cac1-4f80-a489-9e9fb9c730b5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cd0e43f7-cac1-4f80-a489-9e9fb9c730b5\") " pod="openstack/nova-scheduler-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.400202 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/353cb13b-1538-496c-9a26-47f90b784aa4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"353cb13b-1538-496c-9a26-47f90b784aa4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.400254 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd0e43f7-cac1-4f80-a489-9e9fb9c730b5-config-data\") pod \"nova-scheduler-0\" (UID: \"cd0e43f7-cac1-4f80-a489-9e9fb9c730b5\") " pod="openstack/nova-scheduler-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.400435 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pw8gj\" (UniqueName: \"kubernetes.io/projected/cd0e43f7-cac1-4f80-a489-9e9fb9c730b5-kube-api-access-pw8gj\") pod \"nova-scheduler-0\" (UID: \"cd0e43f7-cac1-4f80-a489-9e9fb9c730b5\") " pod="openstack/nova-scheduler-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.400715 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-678qv\" (UniqueName: \"kubernetes.io/projected/353cb13b-1538-496c-9a26-47f90b784aa4-kube-api-access-678qv\") pod \"nova-cell1-novncproxy-0\" (UID: \"353cb13b-1538-496c-9a26-47f90b784aa4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.400745 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/353cb13b-1538-496c-9a26-47f90b784aa4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"353cb13b-1538-496c-9a26-47f90b784aa4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.401052 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.405694 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd0e43f7-cac1-4f80-a489-9e9fb9c730b5-config-data\") pod \"nova-scheduler-0\" (UID: \"cd0e43f7-cac1-4f80-a489-9e9fb9c730b5\") " pod="openstack/nova-scheduler-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.419371 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd0e43f7-cac1-4f80-a489-9e9fb9c730b5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cd0e43f7-cac1-4f80-a489-9e9fb9c730b5\") " pod="openstack/nova-scheduler-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.422529 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pw8gj\" (UniqueName: \"kubernetes.io/projected/cd0e43f7-cac1-4f80-a489-9e9fb9c730b5-kube-api-access-pw8gj\") pod \"nova-scheduler-0\" (UID: \"cd0e43f7-cac1-4f80-a489-9e9fb9c730b5\") " pod="openstack/nova-scheduler-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.502636 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fc43cc8-61cc-4102-aacc-356966a9b9a5-ovsdbserver-nb\") pod \"dnsmasq-dns-75fb48c489-xpjqw\" (UID: \"7fc43cc8-61cc-4102-aacc-356966a9b9a5\") " pod="openstack/dnsmasq-dns-75fb48c489-xpjqw" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.502729 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fc43cc8-61cc-4102-aacc-356966a9b9a5-config\") pod \"dnsmasq-dns-75fb48c489-xpjqw\" (UID: \"7fc43cc8-61cc-4102-aacc-356966a9b9a5\") " pod="openstack/dnsmasq-dns-75fb48c489-xpjqw" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.502757 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7fc43cc8-61cc-4102-aacc-356966a9b9a5-ovsdbserver-sb\") pod \"dnsmasq-dns-75fb48c489-xpjqw\" (UID: \"7fc43cc8-61cc-4102-aacc-356966a9b9a5\") " pod="openstack/dnsmasq-dns-75fb48c489-xpjqw" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.502794 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/353cb13b-1538-496c-9a26-47f90b784aa4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"353cb13b-1538-496c-9a26-47f90b784aa4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.502837 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fc43cc8-61cc-4102-aacc-356966a9b9a5-dns-svc\") pod \"dnsmasq-dns-75fb48c489-xpjqw\" (UID: \"7fc43cc8-61cc-4102-aacc-356966a9b9a5\") " pod="openstack/dnsmasq-dns-75fb48c489-xpjqw" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.502880 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-678qv\" (UniqueName: \"kubernetes.io/projected/353cb13b-1538-496c-9a26-47f90b784aa4-kube-api-access-678qv\") pod \"nova-cell1-novncproxy-0\" (UID: \"353cb13b-1538-496c-9a26-47f90b784aa4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.502909 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/353cb13b-1538-496c-9a26-47f90b784aa4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"353cb13b-1538-496c-9a26-47f90b784aa4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.502933 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khxkr\" (UniqueName: \"kubernetes.io/projected/7fc43cc8-61cc-4102-aacc-356966a9b9a5-kube-api-access-khxkr\") pod \"dnsmasq-dns-75fb48c489-xpjqw\" (UID: \"7fc43cc8-61cc-4102-aacc-356966a9b9a5\") " pod="openstack/dnsmasq-dns-75fb48c489-xpjqw" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.507419 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/353cb13b-1538-496c-9a26-47f90b784aa4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"353cb13b-1538-496c-9a26-47f90b784aa4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.512759 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/353cb13b-1538-496c-9a26-47f90b784aa4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"353cb13b-1538-496c-9a26-47f90b784aa4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.527876 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-678qv\" (UniqueName: \"kubernetes.io/projected/353cb13b-1538-496c-9a26-47f90b784aa4-kube-api-access-678qv\") pod \"nova-cell1-novncproxy-0\" (UID: \"353cb13b-1538-496c-9a26-47f90b784aa4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.607022 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fc43cc8-61cc-4102-aacc-356966a9b9a5-config\") pod \"dnsmasq-dns-75fb48c489-xpjqw\" (UID: \"7fc43cc8-61cc-4102-aacc-356966a9b9a5\") " pod="openstack/dnsmasq-dns-75fb48c489-xpjqw" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.607067 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7fc43cc8-61cc-4102-aacc-356966a9b9a5-ovsdbserver-sb\") pod \"dnsmasq-dns-75fb48c489-xpjqw\" (UID: \"7fc43cc8-61cc-4102-aacc-356966a9b9a5\") " pod="openstack/dnsmasq-dns-75fb48c489-xpjqw" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.607150 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fc43cc8-61cc-4102-aacc-356966a9b9a5-dns-svc\") pod \"dnsmasq-dns-75fb48c489-xpjqw\" (UID: \"7fc43cc8-61cc-4102-aacc-356966a9b9a5\") " pod="openstack/dnsmasq-dns-75fb48c489-xpjqw" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.607204 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khxkr\" (UniqueName: \"kubernetes.io/projected/7fc43cc8-61cc-4102-aacc-356966a9b9a5-kube-api-access-khxkr\") pod \"dnsmasq-dns-75fb48c489-xpjqw\" (UID: \"7fc43cc8-61cc-4102-aacc-356966a9b9a5\") " pod="openstack/dnsmasq-dns-75fb48c489-xpjqw" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.607231 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fc43cc8-61cc-4102-aacc-356966a9b9a5-ovsdbserver-nb\") pod \"dnsmasq-dns-75fb48c489-xpjqw\" (UID: \"7fc43cc8-61cc-4102-aacc-356966a9b9a5\") " pod="openstack/dnsmasq-dns-75fb48c489-xpjqw" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.608122 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fc43cc8-61cc-4102-aacc-356966a9b9a5-ovsdbserver-nb\") pod \"dnsmasq-dns-75fb48c489-xpjqw\" (UID: \"7fc43cc8-61cc-4102-aacc-356966a9b9a5\") " pod="openstack/dnsmasq-dns-75fb48c489-xpjqw" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.608232 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7fc43cc8-61cc-4102-aacc-356966a9b9a5-ovsdbserver-sb\") pod \"dnsmasq-dns-75fb48c489-xpjqw\" (UID: \"7fc43cc8-61cc-4102-aacc-356966a9b9a5\") " pod="openstack/dnsmasq-dns-75fb48c489-xpjqw" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.608438 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fc43cc8-61cc-4102-aacc-356966a9b9a5-dns-svc\") pod \"dnsmasq-dns-75fb48c489-xpjqw\" (UID: \"7fc43cc8-61cc-4102-aacc-356966a9b9a5\") " pod="openstack/dnsmasq-dns-75fb48c489-xpjqw" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.608751 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fc43cc8-61cc-4102-aacc-356966a9b9a5-config\") pod \"dnsmasq-dns-75fb48c489-xpjqw\" (UID: \"7fc43cc8-61cc-4102-aacc-356966a9b9a5\") " pod="openstack/dnsmasq-dns-75fb48c489-xpjqw" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.642905 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khxkr\" (UniqueName: \"kubernetes.io/projected/7fc43cc8-61cc-4102-aacc-356966a9b9a5-kube-api-access-khxkr\") pod \"dnsmasq-dns-75fb48c489-xpjqw\" (UID: \"7fc43cc8-61cc-4102-aacc-356966a9b9a5\") " pod="openstack/dnsmasq-dns-75fb48c489-xpjqw" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.666271 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.681575 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.699509 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75fb48c489-xpjqw" Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.836945 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-qltzt"] Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.857975 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 14:05:19 crc kubenswrapper[4959]: I1007 14:05:19.918801 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 14:05:19 crc kubenswrapper[4959]: W1007 14:05:19.993402 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1fbef70d_f88f_4a9b_bbfd_cabbba6ab107.slice/crio-6f4f0e8d7f0d2b1b6316cb45476dc9fd3fc8288db131e12a4f641b184f8eae11 WatchSource:0}: Error finding container 6f4f0e8d7f0d2b1b6316cb45476dc9fd3fc8288db131e12a4f641b184f8eae11: Status 404 returned error can't find the container with id 6f4f0e8d7f0d2b1b6316cb45476dc9fd3fc8288db131e12a4f641b184f8eae11 Oct 07 14:05:19 crc kubenswrapper[4959]: W1007 14:05:19.999152 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6cbd6c9c_65e4_46d3_85cb_8e20f8b4e6c4.slice/crio-2874adb343f47ca7cda30d3bff15f8aa8965556d3584506f2ec8d36a2e1a1605 WatchSource:0}: Error finding container 2874adb343f47ca7cda30d3bff15f8aa8965556d3584506f2ec8d36a2e1a1605: Status 404 returned error can't find the container with id 2874adb343f47ca7cda30d3bff15f8aa8965556d3584506f2ec8d36a2e1a1605 Oct 07 14:05:20 crc kubenswrapper[4959]: I1007 14:05:20.054673 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-bkbj4"] Oct 07 14:05:20 crc kubenswrapper[4959]: I1007 14:05:20.056342 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-bkbj4" Oct 07 14:05:20 crc kubenswrapper[4959]: I1007 14:05:20.060898 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 07 14:05:20 crc kubenswrapper[4959]: I1007 14:05:20.061088 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 07 14:05:20 crc kubenswrapper[4959]: I1007 14:05:20.074273 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qltzt" event={"ID":"1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f","Type":"ContainerStarted","Data":"4cf5f079de8c489e32f83dead3bfdf5c4dcf160c786b6b8fb08a826efe8fac71"} Oct 07 14:05:20 crc kubenswrapper[4959]: I1007 14:05:20.078925 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1fbef70d-f88f-4a9b-bbfd-cabbba6ab107","Type":"ContainerStarted","Data":"6f4f0e8d7f0d2b1b6316cb45476dc9fd3fc8288db131e12a4f641b184f8eae11"} Oct 07 14:05:20 crc kubenswrapper[4959]: I1007 14:05:20.080005 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4","Type":"ContainerStarted","Data":"2874adb343f47ca7cda30d3bff15f8aa8965556d3584506f2ec8d36a2e1a1605"} Oct 07 14:05:20 crc kubenswrapper[4959]: I1007 14:05:20.085278 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-bkbj4"] Oct 07 14:05:20 crc kubenswrapper[4959]: I1007 14:05:20.135048 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/859a162a-5f11-416b-8156-bf39ba6bb53e-config-data\") pod \"nova-cell1-conductor-db-sync-bkbj4\" (UID: \"859a162a-5f11-416b-8156-bf39ba6bb53e\") " pod="openstack/nova-cell1-conductor-db-sync-bkbj4" Oct 07 14:05:20 crc kubenswrapper[4959]: I1007 14:05:20.135205 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/859a162a-5f11-416b-8156-bf39ba6bb53e-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-bkbj4\" (UID: \"859a162a-5f11-416b-8156-bf39ba6bb53e\") " pod="openstack/nova-cell1-conductor-db-sync-bkbj4" Oct 07 14:05:20 crc kubenswrapper[4959]: I1007 14:05:20.135297 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mqsx\" (UniqueName: \"kubernetes.io/projected/859a162a-5f11-416b-8156-bf39ba6bb53e-kube-api-access-5mqsx\") pod \"nova-cell1-conductor-db-sync-bkbj4\" (UID: \"859a162a-5f11-416b-8156-bf39ba6bb53e\") " pod="openstack/nova-cell1-conductor-db-sync-bkbj4" Oct 07 14:05:20 crc kubenswrapper[4959]: I1007 14:05:20.135429 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/859a162a-5f11-416b-8156-bf39ba6bb53e-scripts\") pod \"nova-cell1-conductor-db-sync-bkbj4\" (UID: \"859a162a-5f11-416b-8156-bf39ba6bb53e\") " pod="openstack/nova-cell1-conductor-db-sync-bkbj4" Oct 07 14:05:20 crc kubenswrapper[4959]: I1007 14:05:20.236827 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/859a162a-5f11-416b-8156-bf39ba6bb53e-scripts\") pod \"nova-cell1-conductor-db-sync-bkbj4\" (UID: \"859a162a-5f11-416b-8156-bf39ba6bb53e\") " pod="openstack/nova-cell1-conductor-db-sync-bkbj4" Oct 07 14:05:20 crc kubenswrapper[4959]: I1007 14:05:20.237174 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/859a162a-5f11-416b-8156-bf39ba6bb53e-config-data\") pod \"nova-cell1-conductor-db-sync-bkbj4\" (UID: \"859a162a-5f11-416b-8156-bf39ba6bb53e\") " pod="openstack/nova-cell1-conductor-db-sync-bkbj4" Oct 07 14:05:20 crc kubenswrapper[4959]: I1007 14:05:20.237392 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/859a162a-5f11-416b-8156-bf39ba6bb53e-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-bkbj4\" (UID: \"859a162a-5f11-416b-8156-bf39ba6bb53e\") " pod="openstack/nova-cell1-conductor-db-sync-bkbj4" Oct 07 14:05:20 crc kubenswrapper[4959]: I1007 14:05:20.237432 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mqsx\" (UniqueName: \"kubernetes.io/projected/859a162a-5f11-416b-8156-bf39ba6bb53e-kube-api-access-5mqsx\") pod \"nova-cell1-conductor-db-sync-bkbj4\" (UID: \"859a162a-5f11-416b-8156-bf39ba6bb53e\") " pod="openstack/nova-cell1-conductor-db-sync-bkbj4" Oct 07 14:05:20 crc kubenswrapper[4959]: I1007 14:05:20.245382 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/859a162a-5f11-416b-8156-bf39ba6bb53e-scripts\") pod \"nova-cell1-conductor-db-sync-bkbj4\" (UID: \"859a162a-5f11-416b-8156-bf39ba6bb53e\") " pod="openstack/nova-cell1-conductor-db-sync-bkbj4" Oct 07 14:05:20 crc kubenswrapper[4959]: I1007 14:05:20.248003 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/859a162a-5f11-416b-8156-bf39ba6bb53e-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-bkbj4\" (UID: \"859a162a-5f11-416b-8156-bf39ba6bb53e\") " pod="openstack/nova-cell1-conductor-db-sync-bkbj4" Oct 07 14:05:20 crc kubenswrapper[4959]: I1007 14:05:20.252173 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/859a162a-5f11-416b-8156-bf39ba6bb53e-config-data\") pod \"nova-cell1-conductor-db-sync-bkbj4\" (UID: \"859a162a-5f11-416b-8156-bf39ba6bb53e\") " pod="openstack/nova-cell1-conductor-db-sync-bkbj4" Oct 07 14:05:20 crc kubenswrapper[4959]: I1007 14:05:20.261781 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mqsx\" (UniqueName: \"kubernetes.io/projected/859a162a-5f11-416b-8156-bf39ba6bb53e-kube-api-access-5mqsx\") pod \"nova-cell1-conductor-db-sync-bkbj4\" (UID: \"859a162a-5f11-416b-8156-bf39ba6bb53e\") " pod="openstack/nova-cell1-conductor-db-sync-bkbj4" Oct 07 14:05:20 crc kubenswrapper[4959]: I1007 14:05:20.327375 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 14:05:20 crc kubenswrapper[4959]: W1007 14:05:20.330284 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd0e43f7_cac1_4f80_a489_9e9fb9c730b5.slice/crio-1b09cae32c47905184c09f1d54a92634f54f6e7d5fb7e340f8518531a54f689d WatchSource:0}: Error finding container 1b09cae32c47905184c09f1d54a92634f54f6e7d5fb7e340f8518531a54f689d: Status 404 returned error can't find the container with id 1b09cae32c47905184c09f1d54a92634f54f6e7d5fb7e340f8518531a54f689d Oct 07 14:05:20 crc kubenswrapper[4959]: I1007 14:05:20.335329 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75fb48c489-xpjqw"] Oct 07 14:05:20 crc kubenswrapper[4959]: I1007 14:05:20.375068 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-bkbj4" Oct 07 14:05:20 crc kubenswrapper[4959]: I1007 14:05:20.449694 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 14:05:20 crc kubenswrapper[4959]: W1007 14:05:20.453627 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod353cb13b_1538_496c_9a26_47f90b784aa4.slice/crio-8138d5710fb875595ba16f5ed50ccd91512a4e403ad3f7fba85c2eb2a4f71379 WatchSource:0}: Error finding container 8138d5710fb875595ba16f5ed50ccd91512a4e403ad3f7fba85c2eb2a4f71379: Status 404 returned error can't find the container with id 8138d5710fb875595ba16f5ed50ccd91512a4e403ad3f7fba85c2eb2a4f71379 Oct 07 14:05:20 crc kubenswrapper[4959]: I1007 14:05:20.821879 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-bkbj4"] Oct 07 14:05:20 crc kubenswrapper[4959]: W1007 14:05:20.826606 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod859a162a_5f11_416b_8156_bf39ba6bb53e.slice/crio-a047fada166235047b2fef2479ec69d9e0766cd4b7c86703c705a7f0b2c77c3d WatchSource:0}: Error finding container a047fada166235047b2fef2479ec69d9e0766cd4b7c86703c705a7f0b2c77c3d: Status 404 returned error can't find the container with id a047fada166235047b2fef2479ec69d9e0766cd4b7c86703c705a7f0b2c77c3d Oct 07 14:05:21 crc kubenswrapper[4959]: I1007 14:05:21.096490 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cd0e43f7-cac1-4f80-a489-9e9fb9c730b5","Type":"ContainerStarted","Data":"1b09cae32c47905184c09f1d54a92634f54f6e7d5fb7e340f8518531a54f689d"} Oct 07 14:05:21 crc kubenswrapper[4959]: I1007 14:05:21.100166 4959 generic.go:334] "Generic (PLEG): container finished" podID="7fc43cc8-61cc-4102-aacc-356966a9b9a5" containerID="9105e8d5198acec26c4d4032d4cc9fe24be4b3e3b51be2f34f12c879c005373c" exitCode=0 Oct 07 14:05:21 crc kubenswrapper[4959]: I1007 14:05:21.100251 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75fb48c489-xpjqw" event={"ID":"7fc43cc8-61cc-4102-aacc-356966a9b9a5","Type":"ContainerDied","Data":"9105e8d5198acec26c4d4032d4cc9fe24be4b3e3b51be2f34f12c879c005373c"} Oct 07 14:05:21 crc kubenswrapper[4959]: I1007 14:05:21.100284 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75fb48c489-xpjqw" event={"ID":"7fc43cc8-61cc-4102-aacc-356966a9b9a5","Type":"ContainerStarted","Data":"fd25e7a57f9d5fff59528486ff53d97d7645e6265dd4fe52648bbf1cc594f1a4"} Oct 07 14:05:21 crc kubenswrapper[4959]: I1007 14:05:21.102530 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"353cb13b-1538-496c-9a26-47f90b784aa4","Type":"ContainerStarted","Data":"8138d5710fb875595ba16f5ed50ccd91512a4e403ad3f7fba85c2eb2a4f71379"} Oct 07 14:05:21 crc kubenswrapper[4959]: I1007 14:05:21.105215 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qltzt" event={"ID":"1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f","Type":"ContainerStarted","Data":"4a4fdfe14783d977fa28e9ac9201e6c4e62084c2ea320ec1660feedafe6fbe6b"} Oct 07 14:05:21 crc kubenswrapper[4959]: I1007 14:05:21.107583 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-bkbj4" event={"ID":"859a162a-5f11-416b-8156-bf39ba6bb53e","Type":"ContainerStarted","Data":"d39161cb1f9795c46badb98f2d6b46cebd3cef33dd7e54fd7fe68d89f992726c"} Oct 07 14:05:21 crc kubenswrapper[4959]: I1007 14:05:21.107615 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-bkbj4" event={"ID":"859a162a-5f11-416b-8156-bf39ba6bb53e","Type":"ContainerStarted","Data":"a047fada166235047b2fef2479ec69d9e0766cd4b7c86703c705a7f0b2c77c3d"} Oct 07 14:05:21 crc kubenswrapper[4959]: I1007 14:05:21.144817 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-bkbj4" podStartSLOduration=1.144796564 podStartE2EDuration="1.144796564s" podCreationTimestamp="2025-10-07 14:05:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:05:21.137227104 +0000 UTC m=+1203.220631429" watchObservedRunningTime="2025-10-07 14:05:21.144796564 +0000 UTC m=+1203.228200889" Oct 07 14:05:21 crc kubenswrapper[4959]: I1007 14:05:21.162305 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-qltzt" podStartSLOduration=3.162280172 podStartE2EDuration="3.162280172s" podCreationTimestamp="2025-10-07 14:05:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:05:21.159699457 +0000 UTC m=+1203.243103782" watchObservedRunningTime="2025-10-07 14:05:21.162280172 +0000 UTC m=+1203.245684497" Oct 07 14:05:22 crc kubenswrapper[4959]: I1007 14:05:22.120315 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75fb48c489-xpjqw" event={"ID":"7fc43cc8-61cc-4102-aacc-356966a9b9a5","Type":"ContainerStarted","Data":"037d18a203529a0c452135fb14b887cb6170bcdf0231d57206c64653e800d31e"} Oct 07 14:05:23 crc kubenswrapper[4959]: I1007 14:05:23.072347 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75fb48c489-xpjqw" podStartSLOduration=4.072328812 podStartE2EDuration="4.072328812s" podCreationTimestamp="2025-10-07 14:05:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:05:22.151024093 +0000 UTC m=+1204.234428408" watchObservedRunningTime="2025-10-07 14:05:23.072328812 +0000 UTC m=+1205.155733127" Oct 07 14:05:23 crc kubenswrapper[4959]: I1007 14:05:23.077655 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 14:05:23 crc kubenswrapper[4959]: I1007 14:05:23.090595 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 14:05:23 crc kubenswrapper[4959]: I1007 14:05:23.128229 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75fb48c489-xpjqw" Oct 07 14:05:24 crc kubenswrapper[4959]: I1007 14:05:24.145750 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"353cb13b-1538-496c-9a26-47f90b784aa4","Type":"ContainerStarted","Data":"b2f9ae43c476cc206537b7de7cdda2fad5b2b7a01f98e0d9c4f2ec7e29936e54"} Oct 07 14:05:24 crc kubenswrapper[4959]: I1007 14:05:24.145858 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="353cb13b-1538-496c-9a26-47f90b784aa4" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://b2f9ae43c476cc206537b7de7cdda2fad5b2b7a01f98e0d9c4f2ec7e29936e54" gracePeriod=30 Oct 07 14:05:24 crc kubenswrapper[4959]: I1007 14:05:24.151223 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1fbef70d-f88f-4a9b-bbfd-cabbba6ab107","Type":"ContainerStarted","Data":"c72e1ceefde275c1043f6aa2ce2b4e66addae8eb914dad971b7ade250648dff1"} Oct 07 14:05:24 crc kubenswrapper[4959]: I1007 14:05:24.153108 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4","Type":"ContainerStarted","Data":"2b1c3e79a528e312b927723b1c9b7dc3fbab7377f8050c40368f7ffefd84eadb"} Oct 07 14:05:24 crc kubenswrapper[4959]: I1007 14:05:24.168915 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=1.945175618 podStartE2EDuration="5.168897215s" podCreationTimestamp="2025-10-07 14:05:19 +0000 UTC" firstStartedPulling="2025-10-07 14:05:20.457113632 +0000 UTC m=+1202.540517957" lastFinishedPulling="2025-10-07 14:05:23.680835229 +0000 UTC m=+1205.764239554" observedRunningTime="2025-10-07 14:05:24.163651774 +0000 UTC m=+1206.247056099" watchObservedRunningTime="2025-10-07 14:05:24.168897215 +0000 UTC m=+1206.252301540" Oct 07 14:05:24 crc kubenswrapper[4959]: I1007 14:05:24.170506 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cd0e43f7-cac1-4f80-a489-9e9fb9c730b5","Type":"ContainerStarted","Data":"18031b1a9be033a4e4fadbf95c756c4a2fa3c53951a4833acd26f02e2ea91e10"} Oct 07 14:05:24 crc kubenswrapper[4959]: I1007 14:05:24.196776 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.851411227 podStartE2EDuration="5.196756844s" podCreationTimestamp="2025-10-07 14:05:19 +0000 UTC" firstStartedPulling="2025-10-07 14:05:20.334672082 +0000 UTC m=+1202.418076407" lastFinishedPulling="2025-10-07 14:05:23.680017699 +0000 UTC m=+1205.763422024" observedRunningTime="2025-10-07 14:05:24.191670966 +0000 UTC m=+1206.275075291" watchObservedRunningTime="2025-10-07 14:05:24.196756844 +0000 UTC m=+1206.280161169" Oct 07 14:05:24 crc kubenswrapper[4959]: I1007 14:05:24.667230 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 07 14:05:24 crc kubenswrapper[4959]: I1007 14:05:24.682751 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:25 crc kubenswrapper[4959]: I1007 14:05:25.179887 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1fbef70d-f88f-4a9b-bbfd-cabbba6ab107","Type":"ContainerStarted","Data":"6704b54501cf7fad38d8058c3cab86b4f30bf644b305b76a93a808c9c796dfcb"} Oct 07 14:05:25 crc kubenswrapper[4959]: I1007 14:05:25.185061 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4" containerName="nova-metadata-log" containerID="cri-o://2b1c3e79a528e312b927723b1c9b7dc3fbab7377f8050c40368f7ffefd84eadb" gracePeriod=30 Oct 07 14:05:25 crc kubenswrapper[4959]: I1007 14:05:25.185454 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4","Type":"ContainerStarted","Data":"2dd014c0e30703e1ae2e3f3d6b63a8db2a50762ecac2d69887a6499147fe345f"} Oct 07 14:05:25 crc kubenswrapper[4959]: I1007 14:05:25.185575 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4" containerName="nova-metadata-metadata" containerID="cri-o://2dd014c0e30703e1ae2e3f3d6b63a8db2a50762ecac2d69887a6499147fe345f" gracePeriod=30 Oct 07 14:05:25 crc kubenswrapper[4959]: I1007 14:05:25.214179 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.527665283 podStartE2EDuration="7.214154492s" podCreationTimestamp="2025-10-07 14:05:18 +0000 UTC" firstStartedPulling="2025-10-07 14:05:20.00112151 +0000 UTC m=+1202.084525835" lastFinishedPulling="2025-10-07 14:05:23.687610719 +0000 UTC m=+1205.771015044" observedRunningTime="2025-10-07 14:05:25.204701605 +0000 UTC m=+1207.288105990" watchObservedRunningTime="2025-10-07 14:05:25.214154492 +0000 UTC m=+1207.297558817" Oct 07 14:05:25 crc kubenswrapper[4959]: I1007 14:05:25.237246 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.563766159 podStartE2EDuration="6.237220361s" podCreationTimestamp="2025-10-07 14:05:19 +0000 UTC" firstStartedPulling="2025-10-07 14:05:20.006841133 +0000 UTC m=+1202.090245458" lastFinishedPulling="2025-10-07 14:05:23.680295335 +0000 UTC m=+1205.763699660" observedRunningTime="2025-10-07 14:05:25.224616195 +0000 UTC m=+1207.308020530" watchObservedRunningTime="2025-10-07 14:05:25.237220361 +0000 UTC m=+1207.320624686" Oct 07 14:05:25 crc kubenswrapper[4959]: I1007 14:05:25.767315 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 14:05:25 crc kubenswrapper[4959]: I1007 14:05:25.950093 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4-combined-ca-bundle\") pod \"6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4\" (UID: \"6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4\") " Oct 07 14:05:25 crc kubenswrapper[4959]: I1007 14:05:25.950255 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xk2s\" (UniqueName: \"kubernetes.io/projected/6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4-kube-api-access-4xk2s\") pod \"6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4\" (UID: \"6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4\") " Oct 07 14:05:25 crc kubenswrapper[4959]: I1007 14:05:25.950325 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4-logs\") pod \"6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4\" (UID: \"6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4\") " Oct 07 14:05:25 crc kubenswrapper[4959]: I1007 14:05:25.950376 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4-config-data\") pod \"6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4\" (UID: \"6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4\") " Oct 07 14:05:25 crc kubenswrapper[4959]: I1007 14:05:25.951726 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4-logs" (OuterVolumeSpecName: "logs") pod "6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4" (UID: "6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:05:25 crc kubenswrapper[4959]: I1007 14:05:25.956874 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4-kube-api-access-4xk2s" (OuterVolumeSpecName: "kube-api-access-4xk2s") pod "6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4" (UID: "6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4"). InnerVolumeSpecName "kube-api-access-4xk2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:05:25 crc kubenswrapper[4959]: I1007 14:05:25.985186 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4-config-data" (OuterVolumeSpecName: "config-data") pod "6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4" (UID: "6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:05:25 crc kubenswrapper[4959]: I1007 14:05:25.985624 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4" (UID: "6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.052743 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.052802 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.052814 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xk2s\" (UniqueName: \"kubernetes.io/projected/6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4-kube-api-access-4xk2s\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.052823 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4-logs\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.219586 4959 generic.go:334] "Generic (PLEG): container finished" podID="6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4" containerID="2dd014c0e30703e1ae2e3f3d6b63a8db2a50762ecac2d69887a6499147fe345f" exitCode=0 Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.219622 4959 generic.go:334] "Generic (PLEG): container finished" podID="6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4" containerID="2b1c3e79a528e312b927723b1c9b7dc3fbab7377f8050c40368f7ffefd84eadb" exitCode=143 Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.220674 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.222261 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4","Type":"ContainerDied","Data":"2dd014c0e30703e1ae2e3f3d6b63a8db2a50762ecac2d69887a6499147fe345f"} Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.222304 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4","Type":"ContainerDied","Data":"2b1c3e79a528e312b927723b1c9b7dc3fbab7377f8050c40368f7ffefd84eadb"} Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.222319 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4","Type":"ContainerDied","Data":"2874adb343f47ca7cda30d3bff15f8aa8965556d3584506f2ec8d36a2e1a1605"} Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.222337 4959 scope.go:117] "RemoveContainer" containerID="2dd014c0e30703e1ae2e3f3d6b63a8db2a50762ecac2d69887a6499147fe345f" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.286150 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.295659 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.322239 4959 scope.go:117] "RemoveContainer" containerID="2b1c3e79a528e312b927723b1c9b7dc3fbab7377f8050c40368f7ffefd84eadb" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.326500 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 07 14:05:26 crc kubenswrapper[4959]: E1007 14:05:26.326865 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4" containerName="nova-metadata-log" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.326882 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4" containerName="nova-metadata-log" Oct 07 14:05:26 crc kubenswrapper[4959]: E1007 14:05:26.326896 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4" containerName="nova-metadata-metadata" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.326903 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4" containerName="nova-metadata-metadata" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.327072 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4" containerName="nova-metadata-metadata" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.327092 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4" containerName="nova-metadata-log" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.328004 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.333513 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.334078 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.347365 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.391594 4959 scope.go:117] "RemoveContainer" containerID="2dd014c0e30703e1ae2e3f3d6b63a8db2a50762ecac2d69887a6499147fe345f" Oct 07 14:05:26 crc kubenswrapper[4959]: E1007 14:05:26.396482 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2dd014c0e30703e1ae2e3f3d6b63a8db2a50762ecac2d69887a6499147fe345f\": container with ID starting with 2dd014c0e30703e1ae2e3f3d6b63a8db2a50762ecac2d69887a6499147fe345f not found: ID does not exist" containerID="2dd014c0e30703e1ae2e3f3d6b63a8db2a50762ecac2d69887a6499147fe345f" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.396531 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dd014c0e30703e1ae2e3f3d6b63a8db2a50762ecac2d69887a6499147fe345f"} err="failed to get container status \"2dd014c0e30703e1ae2e3f3d6b63a8db2a50762ecac2d69887a6499147fe345f\": rpc error: code = NotFound desc = could not find container \"2dd014c0e30703e1ae2e3f3d6b63a8db2a50762ecac2d69887a6499147fe345f\": container with ID starting with 2dd014c0e30703e1ae2e3f3d6b63a8db2a50762ecac2d69887a6499147fe345f not found: ID does not exist" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.396567 4959 scope.go:117] "RemoveContainer" containerID="2b1c3e79a528e312b927723b1c9b7dc3fbab7377f8050c40368f7ffefd84eadb" Oct 07 14:05:26 crc kubenswrapper[4959]: E1007 14:05:26.398991 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b1c3e79a528e312b927723b1c9b7dc3fbab7377f8050c40368f7ffefd84eadb\": container with ID starting with 2b1c3e79a528e312b927723b1c9b7dc3fbab7377f8050c40368f7ffefd84eadb not found: ID does not exist" containerID="2b1c3e79a528e312b927723b1c9b7dc3fbab7377f8050c40368f7ffefd84eadb" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.399029 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b1c3e79a528e312b927723b1c9b7dc3fbab7377f8050c40368f7ffefd84eadb"} err="failed to get container status \"2b1c3e79a528e312b927723b1c9b7dc3fbab7377f8050c40368f7ffefd84eadb\": rpc error: code = NotFound desc = could not find container \"2b1c3e79a528e312b927723b1c9b7dc3fbab7377f8050c40368f7ffefd84eadb\": container with ID starting with 2b1c3e79a528e312b927723b1c9b7dc3fbab7377f8050c40368f7ffefd84eadb not found: ID does not exist" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.399047 4959 scope.go:117] "RemoveContainer" containerID="2dd014c0e30703e1ae2e3f3d6b63a8db2a50762ecac2d69887a6499147fe345f" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.400062 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dd014c0e30703e1ae2e3f3d6b63a8db2a50762ecac2d69887a6499147fe345f"} err="failed to get container status \"2dd014c0e30703e1ae2e3f3d6b63a8db2a50762ecac2d69887a6499147fe345f\": rpc error: code = NotFound desc = could not find container \"2dd014c0e30703e1ae2e3f3d6b63a8db2a50762ecac2d69887a6499147fe345f\": container with ID starting with 2dd014c0e30703e1ae2e3f3d6b63a8db2a50762ecac2d69887a6499147fe345f not found: ID does not exist" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.400085 4959 scope.go:117] "RemoveContainer" containerID="2b1c3e79a528e312b927723b1c9b7dc3fbab7377f8050c40368f7ffefd84eadb" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.400428 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b1c3e79a528e312b927723b1c9b7dc3fbab7377f8050c40368f7ffefd84eadb"} err="failed to get container status \"2b1c3e79a528e312b927723b1c9b7dc3fbab7377f8050c40368f7ffefd84eadb\": rpc error: code = NotFound desc = could not find container \"2b1c3e79a528e312b927723b1c9b7dc3fbab7377f8050c40368f7ffefd84eadb\": container with ID starting with 2b1c3e79a528e312b927723b1c9b7dc3fbab7377f8050c40368f7ffefd84eadb not found: ID does not exist" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.458299 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0970766f-952a-45ff-88ec-ea8b3478b005-logs\") pod \"nova-metadata-0\" (UID: \"0970766f-952a-45ff-88ec-ea8b3478b005\") " pod="openstack/nova-metadata-0" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.458354 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0970766f-952a-45ff-88ec-ea8b3478b005-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0970766f-952a-45ff-88ec-ea8b3478b005\") " pod="openstack/nova-metadata-0" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.458508 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0970766f-952a-45ff-88ec-ea8b3478b005-config-data\") pod \"nova-metadata-0\" (UID: \"0970766f-952a-45ff-88ec-ea8b3478b005\") " pod="openstack/nova-metadata-0" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.458634 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6nb5\" (UniqueName: \"kubernetes.io/projected/0970766f-952a-45ff-88ec-ea8b3478b005-kube-api-access-c6nb5\") pod \"nova-metadata-0\" (UID: \"0970766f-952a-45ff-88ec-ea8b3478b005\") " pod="openstack/nova-metadata-0" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.458843 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0970766f-952a-45ff-88ec-ea8b3478b005-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0970766f-952a-45ff-88ec-ea8b3478b005\") " pod="openstack/nova-metadata-0" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.560981 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0970766f-952a-45ff-88ec-ea8b3478b005-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0970766f-952a-45ff-88ec-ea8b3478b005\") " pod="openstack/nova-metadata-0" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.561063 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0970766f-952a-45ff-88ec-ea8b3478b005-logs\") pod \"nova-metadata-0\" (UID: \"0970766f-952a-45ff-88ec-ea8b3478b005\") " pod="openstack/nova-metadata-0" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.561119 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0970766f-952a-45ff-88ec-ea8b3478b005-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0970766f-952a-45ff-88ec-ea8b3478b005\") " pod="openstack/nova-metadata-0" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.561202 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0970766f-952a-45ff-88ec-ea8b3478b005-config-data\") pod \"nova-metadata-0\" (UID: \"0970766f-952a-45ff-88ec-ea8b3478b005\") " pod="openstack/nova-metadata-0" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.561235 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6nb5\" (UniqueName: \"kubernetes.io/projected/0970766f-952a-45ff-88ec-ea8b3478b005-kube-api-access-c6nb5\") pod \"nova-metadata-0\" (UID: \"0970766f-952a-45ff-88ec-ea8b3478b005\") " pod="openstack/nova-metadata-0" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.562080 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0970766f-952a-45ff-88ec-ea8b3478b005-logs\") pod \"nova-metadata-0\" (UID: \"0970766f-952a-45ff-88ec-ea8b3478b005\") " pod="openstack/nova-metadata-0" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.571013 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0970766f-952a-45ff-88ec-ea8b3478b005-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0970766f-952a-45ff-88ec-ea8b3478b005\") " pod="openstack/nova-metadata-0" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.575384 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0970766f-952a-45ff-88ec-ea8b3478b005-config-data\") pod \"nova-metadata-0\" (UID: \"0970766f-952a-45ff-88ec-ea8b3478b005\") " pod="openstack/nova-metadata-0" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.577997 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6nb5\" (UniqueName: \"kubernetes.io/projected/0970766f-952a-45ff-88ec-ea8b3478b005-kube-api-access-c6nb5\") pod \"nova-metadata-0\" (UID: \"0970766f-952a-45ff-88ec-ea8b3478b005\") " pod="openstack/nova-metadata-0" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.581230 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0970766f-952a-45ff-88ec-ea8b3478b005-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0970766f-952a-45ff-88ec-ea8b3478b005\") " pod="openstack/nova-metadata-0" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.667471 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4" path="/var/lib/kubelet/pods/6cbd6c9c-65e4-46d3-85cb-8e20f8b4e6c4/volumes" Oct 07 14:05:26 crc kubenswrapper[4959]: I1007 14:05:26.672945 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 14:05:27 crc kubenswrapper[4959]: I1007 14:05:27.146980 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 14:05:27 crc kubenswrapper[4959]: W1007 14:05:27.154046 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0970766f_952a_45ff_88ec_ea8b3478b005.slice/crio-fa8c33316f92bd60256c507133c51ff8025e78d8e416f53ca47029a6835ea0b8 WatchSource:0}: Error finding container fa8c33316f92bd60256c507133c51ff8025e78d8e416f53ca47029a6835ea0b8: Status 404 returned error can't find the container with id fa8c33316f92bd60256c507133c51ff8025e78d8e416f53ca47029a6835ea0b8 Oct 07 14:05:27 crc kubenswrapper[4959]: I1007 14:05:27.236854 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0970766f-952a-45ff-88ec-ea8b3478b005","Type":"ContainerStarted","Data":"fa8c33316f92bd60256c507133c51ff8025e78d8e416f53ca47029a6835ea0b8"} Oct 07 14:05:28 crc kubenswrapper[4959]: I1007 14:05:28.244874 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0970766f-952a-45ff-88ec-ea8b3478b005","Type":"ContainerStarted","Data":"bf2a78cd679645e1fcc023456a4fd1433e5d77ac007b8efa48cc7bad049090e2"} Oct 07 14:05:28 crc kubenswrapper[4959]: I1007 14:05:28.245281 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0970766f-952a-45ff-88ec-ea8b3478b005","Type":"ContainerStarted","Data":"dadb0c1553cad1c9b0fc93dfc244c70563d208178d27f33a1552b5860faa5c91"} Oct 07 14:05:28 crc kubenswrapper[4959]: I1007 14:05:28.246810 4959 generic.go:334] "Generic (PLEG): container finished" podID="1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f" containerID="4a4fdfe14783d977fa28e9ac9201e6c4e62084c2ea320ec1660feedafe6fbe6b" exitCode=0 Oct 07 14:05:28 crc kubenswrapper[4959]: I1007 14:05:28.246853 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qltzt" event={"ID":"1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f","Type":"ContainerDied","Data":"4a4fdfe14783d977fa28e9ac9201e6c4e62084c2ea320ec1660feedafe6fbe6b"} Oct 07 14:05:28 crc kubenswrapper[4959]: I1007 14:05:28.277391 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.277372005 podStartE2EDuration="2.277372005s" podCreationTimestamp="2025-10-07 14:05:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:05:28.268065701 +0000 UTC m=+1210.351470036" watchObservedRunningTime="2025-10-07 14:05:28.277372005 +0000 UTC m=+1210.360776330" Oct 07 14:05:29 crc kubenswrapper[4959]: I1007 14:05:29.342449 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 07 14:05:29 crc kubenswrapper[4959]: I1007 14:05:29.342498 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 07 14:05:29 crc kubenswrapper[4959]: I1007 14:05:29.595921 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qltzt" Oct 07 14:05:29 crc kubenswrapper[4959]: I1007 14:05:29.637918 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f-config-data\") pod \"1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f\" (UID: \"1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f\") " Oct 07 14:05:29 crc kubenswrapper[4959]: I1007 14:05:29.638268 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f-combined-ca-bundle\") pod \"1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f\" (UID: \"1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f\") " Oct 07 14:05:29 crc kubenswrapper[4959]: I1007 14:05:29.638389 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f-scripts\") pod \"1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f\" (UID: \"1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f\") " Oct 07 14:05:29 crc kubenswrapper[4959]: I1007 14:05:29.643783 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f-scripts" (OuterVolumeSpecName: "scripts") pod "1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f" (UID: "1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:05:29 crc kubenswrapper[4959]: I1007 14:05:29.667471 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 07 14:05:29 crc kubenswrapper[4959]: I1007 14:05:29.669081 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f" (UID: "1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:05:29 crc kubenswrapper[4959]: I1007 14:05:29.679367 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f-config-data" (OuterVolumeSpecName: "config-data") pod "1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f" (UID: "1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:05:29 crc kubenswrapper[4959]: I1007 14:05:29.700705 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 07 14:05:29 crc kubenswrapper[4959]: I1007 14:05:29.701902 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-75fb48c489-xpjqw" Oct 07 14:05:29 crc kubenswrapper[4959]: I1007 14:05:29.740580 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvzzr\" (UniqueName: \"kubernetes.io/projected/1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f-kube-api-access-lvzzr\") pod \"1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f\" (UID: \"1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f\") " Oct 07 14:05:29 crc kubenswrapper[4959]: I1007 14:05:29.744440 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f-kube-api-access-lvzzr" (OuterVolumeSpecName: "kube-api-access-lvzzr") pod "1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f" (UID: "1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f"). InnerVolumeSpecName "kube-api-access-lvzzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:05:29 crc kubenswrapper[4959]: I1007 14:05:29.747761 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:29 crc kubenswrapper[4959]: I1007 14:05:29.747795 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvzzr\" (UniqueName: \"kubernetes.io/projected/1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f-kube-api-access-lvzzr\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:29 crc kubenswrapper[4959]: I1007 14:05:29.747809 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:29 crc kubenswrapper[4959]: I1007 14:05:29.747821 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:29 crc kubenswrapper[4959]: I1007 14:05:29.778668 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bdc9d6cdc-lzb78"] Oct 07 14:05:29 crc kubenswrapper[4959]: I1007 14:05:29.778925 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7bdc9d6cdc-lzb78" podUID="0789e607-aefa-4c77-9831-050c617faba4" containerName="dnsmasq-dns" containerID="cri-o://f0ebbb837a9273412f30373a4ace9cf1c9ceeeaa6b54fe352f2f2ab03ec94893" gracePeriod=10 Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.181734 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdc9d6cdc-lzb78" Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.255727 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0789e607-aefa-4c77-9831-050c617faba4-ovsdbserver-nb\") pod \"0789e607-aefa-4c77-9831-050c617faba4\" (UID: \"0789e607-aefa-4c77-9831-050c617faba4\") " Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.255844 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrnh5\" (UniqueName: \"kubernetes.io/projected/0789e607-aefa-4c77-9831-050c617faba4-kube-api-access-jrnh5\") pod \"0789e607-aefa-4c77-9831-050c617faba4\" (UID: \"0789e607-aefa-4c77-9831-050c617faba4\") " Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.255910 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0789e607-aefa-4c77-9831-050c617faba4-config\") pod \"0789e607-aefa-4c77-9831-050c617faba4\" (UID: \"0789e607-aefa-4c77-9831-050c617faba4\") " Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.255978 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0789e607-aefa-4c77-9831-050c617faba4-ovsdbserver-sb\") pod \"0789e607-aefa-4c77-9831-050c617faba4\" (UID: \"0789e607-aefa-4c77-9831-050c617faba4\") " Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.256038 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0789e607-aefa-4c77-9831-050c617faba4-dns-svc\") pod \"0789e607-aefa-4c77-9831-050c617faba4\" (UID: \"0789e607-aefa-4c77-9831-050c617faba4\") " Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.272358 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0789e607-aefa-4c77-9831-050c617faba4-kube-api-access-jrnh5" (OuterVolumeSpecName: "kube-api-access-jrnh5") pod "0789e607-aefa-4c77-9831-050c617faba4" (UID: "0789e607-aefa-4c77-9831-050c617faba4"). InnerVolumeSpecName "kube-api-access-jrnh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.302582 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qltzt" Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.302686 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qltzt" event={"ID":"1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f","Type":"ContainerDied","Data":"4cf5f079de8c489e32f83dead3bfdf5c4dcf160c786b6b8fb08a826efe8fac71"} Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.302727 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4cf5f079de8c489e32f83dead3bfdf5c4dcf160c786b6b8fb08a826efe8fac71" Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.330942 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0789e607-aefa-4c77-9831-050c617faba4-config" (OuterVolumeSpecName: "config") pod "0789e607-aefa-4c77-9831-050c617faba4" (UID: "0789e607-aefa-4c77-9831-050c617faba4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.358956 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0789e607-aefa-4c77-9831-050c617faba4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0789e607-aefa-4c77-9831-050c617faba4" (UID: "0789e607-aefa-4c77-9831-050c617faba4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.360124 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrnh5\" (UniqueName: \"kubernetes.io/projected/0789e607-aefa-4c77-9831-050c617faba4-kube-api-access-jrnh5\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.360139 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0789e607-aefa-4c77-9831-050c617faba4-config\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.360148 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0789e607-aefa-4c77-9831-050c617faba4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.368524 4959 generic.go:334] "Generic (PLEG): container finished" podID="0789e607-aefa-4c77-9831-050c617faba4" containerID="f0ebbb837a9273412f30373a4ace9cf1c9ceeeaa6b54fe352f2f2ab03ec94893" exitCode=0 Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.369419 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdc9d6cdc-lzb78" Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.369829 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bdc9d6cdc-lzb78" event={"ID":"0789e607-aefa-4c77-9831-050c617faba4","Type":"ContainerDied","Data":"f0ebbb837a9273412f30373a4ace9cf1c9ceeeaa6b54fe352f2f2ab03ec94893"} Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.369850 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bdc9d6cdc-lzb78" event={"ID":"0789e607-aefa-4c77-9831-050c617faba4","Type":"ContainerDied","Data":"f8e6595792d7d30c10249576b5e934172a40ff5551131d27846c7aff68c032f7"} Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.369866 4959 scope.go:117] "RemoveContainer" containerID="f0ebbb837a9273412f30373a4ace9cf1c9ceeeaa6b54fe352f2f2ab03ec94893" Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.381728 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0789e607-aefa-4c77-9831-050c617faba4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0789e607-aefa-4c77-9831-050c617faba4" (UID: "0789e607-aefa-4c77-9831-050c617faba4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.408932 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0789e607-aefa-4c77-9831-050c617faba4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0789e607-aefa-4c77-9831-050c617faba4" (UID: "0789e607-aefa-4c77-9831-050c617faba4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.431673 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1fbef70d-f88f-4a9b-bbfd-cabbba6ab107" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.177:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.431975 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1fbef70d-f88f-4a9b-bbfd-cabbba6ab107" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.177:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.440808 4959 scope.go:117] "RemoveContainer" containerID="5e7ab95dce143648aaba21fda226685f08f198b1d25b25dbe727037e8ee0ee20" Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.447449 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.464377 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0789e607-aefa-4c77-9831-050c617faba4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.464600 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0789e607-aefa-4c77-9831-050c617faba4-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.480057 4959 scope.go:117] "RemoveContainer" containerID="f0ebbb837a9273412f30373a4ace9cf1c9ceeeaa6b54fe352f2f2ab03ec94893" Oct 07 14:05:30 crc kubenswrapper[4959]: E1007 14:05:30.480498 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0ebbb837a9273412f30373a4ace9cf1c9ceeeaa6b54fe352f2f2ab03ec94893\": container with ID starting with f0ebbb837a9273412f30373a4ace9cf1c9ceeeaa6b54fe352f2f2ab03ec94893 not found: ID does not exist" containerID="f0ebbb837a9273412f30373a4ace9cf1c9ceeeaa6b54fe352f2f2ab03ec94893" Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.480540 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0ebbb837a9273412f30373a4ace9cf1c9ceeeaa6b54fe352f2f2ab03ec94893"} err="failed to get container status \"f0ebbb837a9273412f30373a4ace9cf1c9ceeeaa6b54fe352f2f2ab03ec94893\": rpc error: code = NotFound desc = could not find container \"f0ebbb837a9273412f30373a4ace9cf1c9ceeeaa6b54fe352f2f2ab03ec94893\": container with ID starting with f0ebbb837a9273412f30373a4ace9cf1c9ceeeaa6b54fe352f2f2ab03ec94893 not found: ID does not exist" Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.480574 4959 scope.go:117] "RemoveContainer" containerID="5e7ab95dce143648aaba21fda226685f08f198b1d25b25dbe727037e8ee0ee20" Oct 07 14:05:30 crc kubenswrapper[4959]: E1007 14:05:30.481429 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e7ab95dce143648aaba21fda226685f08f198b1d25b25dbe727037e8ee0ee20\": container with ID starting with 5e7ab95dce143648aaba21fda226685f08f198b1d25b25dbe727037e8ee0ee20 not found: ID does not exist" containerID="5e7ab95dce143648aaba21fda226685f08f198b1d25b25dbe727037e8ee0ee20" Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.481559 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e7ab95dce143648aaba21fda226685f08f198b1d25b25dbe727037e8ee0ee20"} err="failed to get container status \"5e7ab95dce143648aaba21fda226685f08f198b1d25b25dbe727037e8ee0ee20\": rpc error: code = NotFound desc = could not find container \"5e7ab95dce143648aaba21fda226685f08f198b1d25b25dbe727037e8ee0ee20\": container with ID starting with 5e7ab95dce143648aaba21fda226685f08f198b1d25b25dbe727037e8ee0ee20 not found: ID does not exist" Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.583348 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.583982 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1fbef70d-f88f-4a9b-bbfd-cabbba6ab107" containerName="nova-api-api" containerID="cri-o://6704b54501cf7fad38d8058c3cab86b4f30bf644b305b76a93a808c9c796dfcb" gracePeriod=30 Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.583622 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1fbef70d-f88f-4a9b-bbfd-cabbba6ab107" containerName="nova-api-log" containerID="cri-o://c72e1ceefde275c1043f6aa2ce2b4e66addae8eb914dad971b7ade250648dff1" gracePeriod=30 Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.602269 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.602643 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0970766f-952a-45ff-88ec-ea8b3478b005" containerName="nova-metadata-log" containerID="cri-o://dadb0c1553cad1c9b0fc93dfc244c70563d208178d27f33a1552b5860faa5c91" gracePeriod=30 Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.603146 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0970766f-952a-45ff-88ec-ea8b3478b005" containerName="nova-metadata-metadata" containerID="cri-o://bf2a78cd679645e1fcc023456a4fd1433e5d77ac007b8efa48cc7bad049090e2" gracePeriod=30 Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.691915 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bdc9d6cdc-lzb78"] Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.698616 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bdc9d6cdc-lzb78"] Oct 07 14:05:30 crc kubenswrapper[4959]: I1007 14:05:30.888637 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 14:05:31 crc kubenswrapper[4959]: I1007 14:05:31.381885 4959 generic.go:334] "Generic (PLEG): container finished" podID="0970766f-952a-45ff-88ec-ea8b3478b005" containerID="bf2a78cd679645e1fcc023456a4fd1433e5d77ac007b8efa48cc7bad049090e2" exitCode=0 Oct 07 14:05:31 crc kubenswrapper[4959]: I1007 14:05:31.381916 4959 generic.go:334] "Generic (PLEG): container finished" podID="0970766f-952a-45ff-88ec-ea8b3478b005" containerID="dadb0c1553cad1c9b0fc93dfc244c70563d208178d27f33a1552b5860faa5c91" exitCode=143 Oct 07 14:05:31 crc kubenswrapper[4959]: I1007 14:05:31.381954 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0970766f-952a-45ff-88ec-ea8b3478b005","Type":"ContainerDied","Data":"bf2a78cd679645e1fcc023456a4fd1433e5d77ac007b8efa48cc7bad049090e2"} Oct 07 14:05:31 crc kubenswrapper[4959]: I1007 14:05:31.381979 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0970766f-952a-45ff-88ec-ea8b3478b005","Type":"ContainerDied","Data":"dadb0c1553cad1c9b0fc93dfc244c70563d208178d27f33a1552b5860faa5c91"} Oct 07 14:05:31 crc kubenswrapper[4959]: I1007 14:05:31.385062 4959 generic.go:334] "Generic (PLEG): container finished" podID="1fbef70d-f88f-4a9b-bbfd-cabbba6ab107" containerID="c72e1ceefde275c1043f6aa2ce2b4e66addae8eb914dad971b7ade250648dff1" exitCode=143 Oct 07 14:05:31 crc kubenswrapper[4959]: I1007 14:05:31.385208 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1fbef70d-f88f-4a9b-bbfd-cabbba6ab107","Type":"ContainerDied","Data":"c72e1ceefde275c1043f6aa2ce2b4e66addae8eb914dad971b7ade250648dff1"} Oct 07 14:05:31 crc kubenswrapper[4959]: I1007 14:05:31.566350 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 14:05:31 crc kubenswrapper[4959]: I1007 14:05:31.586790 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0970766f-952a-45ff-88ec-ea8b3478b005-nova-metadata-tls-certs\") pod \"0970766f-952a-45ff-88ec-ea8b3478b005\" (UID: \"0970766f-952a-45ff-88ec-ea8b3478b005\") " Oct 07 14:05:31 crc kubenswrapper[4959]: I1007 14:05:31.586923 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0970766f-952a-45ff-88ec-ea8b3478b005-combined-ca-bundle\") pod \"0970766f-952a-45ff-88ec-ea8b3478b005\" (UID: \"0970766f-952a-45ff-88ec-ea8b3478b005\") " Oct 07 14:05:31 crc kubenswrapper[4959]: I1007 14:05:31.586961 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0970766f-952a-45ff-88ec-ea8b3478b005-logs\") pod \"0970766f-952a-45ff-88ec-ea8b3478b005\" (UID: \"0970766f-952a-45ff-88ec-ea8b3478b005\") " Oct 07 14:05:31 crc kubenswrapper[4959]: I1007 14:05:31.586985 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0970766f-952a-45ff-88ec-ea8b3478b005-config-data\") pod \"0970766f-952a-45ff-88ec-ea8b3478b005\" (UID: \"0970766f-952a-45ff-88ec-ea8b3478b005\") " Oct 07 14:05:31 crc kubenswrapper[4959]: I1007 14:05:31.587041 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6nb5\" (UniqueName: \"kubernetes.io/projected/0970766f-952a-45ff-88ec-ea8b3478b005-kube-api-access-c6nb5\") pod \"0970766f-952a-45ff-88ec-ea8b3478b005\" (UID: \"0970766f-952a-45ff-88ec-ea8b3478b005\") " Oct 07 14:05:31 crc kubenswrapper[4959]: I1007 14:05:31.587255 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0970766f-952a-45ff-88ec-ea8b3478b005-logs" (OuterVolumeSpecName: "logs") pod "0970766f-952a-45ff-88ec-ea8b3478b005" (UID: "0970766f-952a-45ff-88ec-ea8b3478b005"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:05:31 crc kubenswrapper[4959]: I1007 14:05:31.587764 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0970766f-952a-45ff-88ec-ea8b3478b005-logs\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:31 crc kubenswrapper[4959]: I1007 14:05:31.597261 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0970766f-952a-45ff-88ec-ea8b3478b005-kube-api-access-c6nb5" (OuterVolumeSpecName: "kube-api-access-c6nb5") pod "0970766f-952a-45ff-88ec-ea8b3478b005" (UID: "0970766f-952a-45ff-88ec-ea8b3478b005"). InnerVolumeSpecName "kube-api-access-c6nb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:05:31 crc kubenswrapper[4959]: I1007 14:05:31.631431 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0970766f-952a-45ff-88ec-ea8b3478b005-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0970766f-952a-45ff-88ec-ea8b3478b005" (UID: "0970766f-952a-45ff-88ec-ea8b3478b005"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:05:31 crc kubenswrapper[4959]: I1007 14:05:31.636482 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0970766f-952a-45ff-88ec-ea8b3478b005-config-data" (OuterVolumeSpecName: "config-data") pod "0970766f-952a-45ff-88ec-ea8b3478b005" (UID: "0970766f-952a-45ff-88ec-ea8b3478b005"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:05:31 crc kubenswrapper[4959]: I1007 14:05:31.648868 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0970766f-952a-45ff-88ec-ea8b3478b005-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "0970766f-952a-45ff-88ec-ea8b3478b005" (UID: "0970766f-952a-45ff-88ec-ea8b3478b005"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:05:31 crc kubenswrapper[4959]: I1007 14:05:31.690959 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0970766f-952a-45ff-88ec-ea8b3478b005-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:31 crc kubenswrapper[4959]: I1007 14:05:31.691276 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0970766f-952a-45ff-88ec-ea8b3478b005-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:31 crc kubenswrapper[4959]: I1007 14:05:31.691287 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6nb5\" (UniqueName: \"kubernetes.io/projected/0970766f-952a-45ff-88ec-ea8b3478b005-kube-api-access-c6nb5\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:31 crc kubenswrapper[4959]: I1007 14:05:31.691297 4959 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0970766f-952a-45ff-88ec-ea8b3478b005-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.402622 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0970766f-952a-45ff-88ec-ea8b3478b005","Type":"ContainerDied","Data":"fa8c33316f92bd60256c507133c51ff8025e78d8e416f53ca47029a6835ea0b8"} Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.402675 4959 scope.go:117] "RemoveContainer" containerID="bf2a78cd679645e1fcc023456a4fd1433e5d77ac007b8efa48cc7bad049090e2" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.402683 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.406556 4959 generic.go:334] "Generic (PLEG): container finished" podID="859a162a-5f11-416b-8156-bf39ba6bb53e" containerID="d39161cb1f9795c46badb98f2d6b46cebd3cef33dd7e54fd7fe68d89f992726c" exitCode=0 Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.406676 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="cd0e43f7-cac1-4f80-a489-9e9fb9c730b5" containerName="nova-scheduler-scheduler" containerID="cri-o://18031b1a9be033a4e4fadbf95c756c4a2fa3c53951a4833acd26f02e2ea91e10" gracePeriod=30 Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.406668 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-bkbj4" event={"ID":"859a162a-5f11-416b-8156-bf39ba6bb53e","Type":"ContainerDied","Data":"d39161cb1f9795c46badb98f2d6b46cebd3cef33dd7e54fd7fe68d89f992726c"} Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.461597 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.465664 4959 scope.go:117] "RemoveContainer" containerID="dadb0c1553cad1c9b0fc93dfc244c70563d208178d27f33a1552b5860faa5c91" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.471193 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.478120 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 07 14:05:32 crc kubenswrapper[4959]: E1007 14:05:32.478470 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0970766f-952a-45ff-88ec-ea8b3478b005" containerName="nova-metadata-metadata" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.478489 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="0970766f-952a-45ff-88ec-ea8b3478b005" containerName="nova-metadata-metadata" Oct 07 14:05:32 crc kubenswrapper[4959]: E1007 14:05:32.478502 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0789e607-aefa-4c77-9831-050c617faba4" containerName="init" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.478508 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="0789e607-aefa-4c77-9831-050c617faba4" containerName="init" Oct 07 14:05:32 crc kubenswrapper[4959]: E1007 14:05:32.478521 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f" containerName="nova-manage" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.478528 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f" containerName="nova-manage" Oct 07 14:05:32 crc kubenswrapper[4959]: E1007 14:05:32.478537 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0789e607-aefa-4c77-9831-050c617faba4" containerName="dnsmasq-dns" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.478542 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="0789e607-aefa-4c77-9831-050c617faba4" containerName="dnsmasq-dns" Oct 07 14:05:32 crc kubenswrapper[4959]: E1007 14:05:32.478558 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0970766f-952a-45ff-88ec-ea8b3478b005" containerName="nova-metadata-log" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.478563 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="0970766f-952a-45ff-88ec-ea8b3478b005" containerName="nova-metadata-log" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.478744 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f" containerName="nova-manage" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.478765 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="0789e607-aefa-4c77-9831-050c617faba4" containerName="dnsmasq-dns" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.478777 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="0970766f-952a-45ff-88ec-ea8b3478b005" containerName="nova-metadata-metadata" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.478786 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="0970766f-952a-45ff-88ec-ea8b3478b005" containerName="nova-metadata-log" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.480046 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.482147 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.484333 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.495528 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.609837 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f\") " pod="openstack/nova-metadata-0" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.609912 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-logs\") pod \"nova-metadata-0\" (UID: \"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f\") " pod="openstack/nova-metadata-0" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.609950 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f\") " pod="openstack/nova-metadata-0" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.610125 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-config-data\") pod \"nova-metadata-0\" (UID: \"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f\") " pod="openstack/nova-metadata-0" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.610199 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcj89\" (UniqueName: \"kubernetes.io/projected/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-kube-api-access-dcj89\") pod \"nova-metadata-0\" (UID: \"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f\") " pod="openstack/nova-metadata-0" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.664483 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0789e607-aefa-4c77-9831-050c617faba4" path="/var/lib/kubelet/pods/0789e607-aefa-4c77-9831-050c617faba4/volumes" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.665163 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0970766f-952a-45ff-88ec-ea8b3478b005" path="/var/lib/kubelet/pods/0970766f-952a-45ff-88ec-ea8b3478b005/volumes" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.710991 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f\") " pod="openstack/nova-metadata-0" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.711066 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-logs\") pod \"nova-metadata-0\" (UID: \"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f\") " pod="openstack/nova-metadata-0" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.711120 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f\") " pod="openstack/nova-metadata-0" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.711146 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-config-data\") pod \"nova-metadata-0\" (UID: \"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f\") " pod="openstack/nova-metadata-0" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.711163 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcj89\" (UniqueName: \"kubernetes.io/projected/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-kube-api-access-dcj89\") pod \"nova-metadata-0\" (UID: \"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f\") " pod="openstack/nova-metadata-0" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.711518 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-logs\") pod \"nova-metadata-0\" (UID: \"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f\") " pod="openstack/nova-metadata-0" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.715309 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f\") " pod="openstack/nova-metadata-0" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.726880 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f\") " pod="openstack/nova-metadata-0" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.730769 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-config-data\") pod \"nova-metadata-0\" (UID: \"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f\") " pod="openstack/nova-metadata-0" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.731873 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcj89\" (UniqueName: \"kubernetes.io/projected/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-kube-api-access-dcj89\") pod \"nova-metadata-0\" (UID: \"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f\") " pod="openstack/nova-metadata-0" Oct 07 14:05:32 crc kubenswrapper[4959]: I1007 14:05:32.808507 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 14:05:33 crc kubenswrapper[4959]: I1007 14:05:33.405786 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 14:05:33 crc kubenswrapper[4959]: I1007 14:05:33.706433 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-bkbj4" Oct 07 14:05:33 crc kubenswrapper[4959]: I1007 14:05:33.826629 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mqsx\" (UniqueName: \"kubernetes.io/projected/859a162a-5f11-416b-8156-bf39ba6bb53e-kube-api-access-5mqsx\") pod \"859a162a-5f11-416b-8156-bf39ba6bb53e\" (UID: \"859a162a-5f11-416b-8156-bf39ba6bb53e\") " Oct 07 14:05:33 crc kubenswrapper[4959]: I1007 14:05:33.826744 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/859a162a-5f11-416b-8156-bf39ba6bb53e-scripts\") pod \"859a162a-5f11-416b-8156-bf39ba6bb53e\" (UID: \"859a162a-5f11-416b-8156-bf39ba6bb53e\") " Oct 07 14:05:33 crc kubenswrapper[4959]: I1007 14:05:33.826799 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/859a162a-5f11-416b-8156-bf39ba6bb53e-config-data\") pod \"859a162a-5f11-416b-8156-bf39ba6bb53e\" (UID: \"859a162a-5f11-416b-8156-bf39ba6bb53e\") " Oct 07 14:05:33 crc kubenswrapper[4959]: I1007 14:05:33.826854 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/859a162a-5f11-416b-8156-bf39ba6bb53e-combined-ca-bundle\") pod \"859a162a-5f11-416b-8156-bf39ba6bb53e\" (UID: \"859a162a-5f11-416b-8156-bf39ba6bb53e\") " Oct 07 14:05:33 crc kubenswrapper[4959]: I1007 14:05:33.830996 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/859a162a-5f11-416b-8156-bf39ba6bb53e-scripts" (OuterVolumeSpecName: "scripts") pod "859a162a-5f11-416b-8156-bf39ba6bb53e" (UID: "859a162a-5f11-416b-8156-bf39ba6bb53e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:05:33 crc kubenswrapper[4959]: I1007 14:05:33.832606 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/859a162a-5f11-416b-8156-bf39ba6bb53e-kube-api-access-5mqsx" (OuterVolumeSpecName: "kube-api-access-5mqsx") pod "859a162a-5f11-416b-8156-bf39ba6bb53e" (UID: "859a162a-5f11-416b-8156-bf39ba6bb53e"). InnerVolumeSpecName "kube-api-access-5mqsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:05:33 crc kubenswrapper[4959]: I1007 14:05:33.851150 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/859a162a-5f11-416b-8156-bf39ba6bb53e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "859a162a-5f11-416b-8156-bf39ba6bb53e" (UID: "859a162a-5f11-416b-8156-bf39ba6bb53e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:05:33 crc kubenswrapper[4959]: I1007 14:05:33.851357 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/859a162a-5f11-416b-8156-bf39ba6bb53e-config-data" (OuterVolumeSpecName: "config-data") pod "859a162a-5f11-416b-8156-bf39ba6bb53e" (UID: "859a162a-5f11-416b-8156-bf39ba6bb53e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:05:33 crc kubenswrapper[4959]: I1007 14:05:33.928077 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/859a162a-5f11-416b-8156-bf39ba6bb53e-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:33 crc kubenswrapper[4959]: I1007 14:05:33.928128 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/859a162a-5f11-416b-8156-bf39ba6bb53e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:33 crc kubenswrapper[4959]: I1007 14:05:33.928143 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mqsx\" (UniqueName: \"kubernetes.io/projected/859a162a-5f11-416b-8156-bf39ba6bb53e-kube-api-access-5mqsx\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:33 crc kubenswrapper[4959]: I1007 14:05:33.928152 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/859a162a-5f11-416b-8156-bf39ba6bb53e-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:34 crc kubenswrapper[4959]: I1007 14:05:34.427167 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f","Type":"ContainerStarted","Data":"8cd1268f6a16cc19070e1500a6303258aa9ca5e7bfa867d9350c4ac21ee4d585"} Oct 07 14:05:34 crc kubenswrapper[4959]: I1007 14:05:34.427459 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f","Type":"ContainerStarted","Data":"c24cd83763ffe59dcb7a1d47b5f5811d39efff6aeb0689d5ea52a3e526d61ba4"} Oct 07 14:05:34 crc kubenswrapper[4959]: I1007 14:05:34.427476 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f","Type":"ContainerStarted","Data":"e27732ede22f998af89667530ea54c3ca08cfbd63240c3ae61b6f2a23fec75aa"} Oct 07 14:05:34 crc kubenswrapper[4959]: I1007 14:05:34.428938 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-bkbj4" Oct 07 14:05:34 crc kubenswrapper[4959]: I1007 14:05:34.428992 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-bkbj4" event={"ID":"859a162a-5f11-416b-8156-bf39ba6bb53e","Type":"ContainerDied","Data":"a047fada166235047b2fef2479ec69d9e0766cd4b7c86703c705a7f0b2c77c3d"} Oct 07 14:05:34 crc kubenswrapper[4959]: I1007 14:05:34.429037 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a047fada166235047b2fef2479ec69d9e0766cd4b7c86703c705a7f0b2c77c3d" Oct 07 14:05:34 crc kubenswrapper[4959]: I1007 14:05:34.463666 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.463644799 podStartE2EDuration="2.463644799s" podCreationTimestamp="2025-10-07 14:05:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:05:34.450031678 +0000 UTC m=+1216.533436003" watchObservedRunningTime="2025-10-07 14:05:34.463644799 +0000 UTC m=+1216.547049124" Oct 07 14:05:34 crc kubenswrapper[4959]: I1007 14:05:34.530971 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 07 14:05:34 crc kubenswrapper[4959]: E1007 14:05:34.531636 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="859a162a-5f11-416b-8156-bf39ba6bb53e" containerName="nova-cell1-conductor-db-sync" Oct 07 14:05:34 crc kubenswrapper[4959]: I1007 14:05:34.531666 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="859a162a-5f11-416b-8156-bf39ba6bb53e" containerName="nova-cell1-conductor-db-sync" Oct 07 14:05:34 crc kubenswrapper[4959]: I1007 14:05:34.532023 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="859a162a-5f11-416b-8156-bf39ba6bb53e" containerName="nova-cell1-conductor-db-sync" Oct 07 14:05:34 crc kubenswrapper[4959]: I1007 14:05:34.533021 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 07 14:05:34 crc kubenswrapper[4959]: I1007 14:05:34.538056 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 07 14:05:34 crc kubenswrapper[4959]: I1007 14:05:34.541386 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 07 14:05:34 crc kubenswrapper[4959]: I1007 14:05:34.646728 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q668d\" (UniqueName: \"kubernetes.io/projected/850dd238-e869-4484-9d52-3c67dc241120-kube-api-access-q668d\") pod \"nova-cell1-conductor-0\" (UID: \"850dd238-e869-4484-9d52-3c67dc241120\") " pod="openstack/nova-cell1-conductor-0" Oct 07 14:05:34 crc kubenswrapper[4959]: I1007 14:05:34.646786 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/850dd238-e869-4484-9d52-3c67dc241120-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"850dd238-e869-4484-9d52-3c67dc241120\") " pod="openstack/nova-cell1-conductor-0" Oct 07 14:05:34 crc kubenswrapper[4959]: I1007 14:05:34.646950 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/850dd238-e869-4484-9d52-3c67dc241120-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"850dd238-e869-4484-9d52-3c67dc241120\") " pod="openstack/nova-cell1-conductor-0" Oct 07 14:05:34 crc kubenswrapper[4959]: E1007 14:05:34.668610 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="18031b1a9be033a4e4fadbf95c756c4a2fa3c53951a4833acd26f02e2ea91e10" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 07 14:05:34 crc kubenswrapper[4959]: E1007 14:05:34.669694 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="18031b1a9be033a4e4fadbf95c756c4a2fa3c53951a4833acd26f02e2ea91e10" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 07 14:05:34 crc kubenswrapper[4959]: E1007 14:05:34.670924 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="18031b1a9be033a4e4fadbf95c756c4a2fa3c53951a4833acd26f02e2ea91e10" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 07 14:05:34 crc kubenswrapper[4959]: E1007 14:05:34.670958 4959 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="cd0e43f7-cac1-4f80-a489-9e9fb9c730b5" containerName="nova-scheduler-scheduler" Oct 07 14:05:34 crc kubenswrapper[4959]: I1007 14:05:34.749236 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/850dd238-e869-4484-9d52-3c67dc241120-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"850dd238-e869-4484-9d52-3c67dc241120\") " pod="openstack/nova-cell1-conductor-0" Oct 07 14:05:34 crc kubenswrapper[4959]: I1007 14:05:34.749329 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q668d\" (UniqueName: \"kubernetes.io/projected/850dd238-e869-4484-9d52-3c67dc241120-kube-api-access-q668d\") pod \"nova-cell1-conductor-0\" (UID: \"850dd238-e869-4484-9d52-3c67dc241120\") " pod="openstack/nova-cell1-conductor-0" Oct 07 14:05:34 crc kubenswrapper[4959]: I1007 14:05:34.749425 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/850dd238-e869-4484-9d52-3c67dc241120-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"850dd238-e869-4484-9d52-3c67dc241120\") " pod="openstack/nova-cell1-conductor-0" Oct 07 14:05:34 crc kubenswrapper[4959]: I1007 14:05:34.753446 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/850dd238-e869-4484-9d52-3c67dc241120-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"850dd238-e869-4484-9d52-3c67dc241120\") " pod="openstack/nova-cell1-conductor-0" Oct 07 14:05:34 crc kubenswrapper[4959]: I1007 14:05:34.753698 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/850dd238-e869-4484-9d52-3c67dc241120-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"850dd238-e869-4484-9d52-3c67dc241120\") " pod="openstack/nova-cell1-conductor-0" Oct 07 14:05:34 crc kubenswrapper[4959]: I1007 14:05:34.765420 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q668d\" (UniqueName: \"kubernetes.io/projected/850dd238-e869-4484-9d52-3c67dc241120-kube-api-access-q668d\") pod \"nova-cell1-conductor-0\" (UID: \"850dd238-e869-4484-9d52-3c67dc241120\") " pod="openstack/nova-cell1-conductor-0" Oct 07 14:05:34 crc kubenswrapper[4959]: I1007 14:05:34.864619 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 07 14:05:35 crc kubenswrapper[4959]: W1007 14:05:35.278273 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod850dd238_e869_4484_9d52_3c67dc241120.slice/crio-324cdaccd75ee151735a2bf3f9b76faef827e9fb4bbc327770126ecc41db3606 WatchSource:0}: Error finding container 324cdaccd75ee151735a2bf3f9b76faef827e9fb4bbc327770126ecc41db3606: Status 404 returned error can't find the container with id 324cdaccd75ee151735a2bf3f9b76faef827e9fb4bbc327770126ecc41db3606 Oct 07 14:05:35 crc kubenswrapper[4959]: I1007 14:05:35.279244 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 07 14:05:35 crc kubenswrapper[4959]: I1007 14:05:35.444709 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"850dd238-e869-4484-9d52-3c67dc241120","Type":"ContainerStarted","Data":"324cdaccd75ee151735a2bf3f9b76faef827e9fb4bbc327770126ecc41db3606"} Oct 07 14:05:35 crc kubenswrapper[4959]: I1007 14:05:35.869841 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 14:05:35 crc kubenswrapper[4959]: I1007 14:05:35.976347 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd0e43f7-cac1-4f80-a489-9e9fb9c730b5-combined-ca-bundle\") pod \"cd0e43f7-cac1-4f80-a489-9e9fb9c730b5\" (UID: \"cd0e43f7-cac1-4f80-a489-9e9fb9c730b5\") " Oct 07 14:05:35 crc kubenswrapper[4959]: I1007 14:05:35.976488 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd0e43f7-cac1-4f80-a489-9e9fb9c730b5-config-data\") pod \"cd0e43f7-cac1-4f80-a489-9e9fb9c730b5\" (UID: \"cd0e43f7-cac1-4f80-a489-9e9fb9c730b5\") " Oct 07 14:05:35 crc kubenswrapper[4959]: I1007 14:05:35.976747 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pw8gj\" (UniqueName: \"kubernetes.io/projected/cd0e43f7-cac1-4f80-a489-9e9fb9c730b5-kube-api-access-pw8gj\") pod \"cd0e43f7-cac1-4f80-a489-9e9fb9c730b5\" (UID: \"cd0e43f7-cac1-4f80-a489-9e9fb9c730b5\") " Oct 07 14:05:35 crc kubenswrapper[4959]: I1007 14:05:35.982088 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd0e43f7-cac1-4f80-a489-9e9fb9c730b5-kube-api-access-pw8gj" (OuterVolumeSpecName: "kube-api-access-pw8gj") pod "cd0e43f7-cac1-4f80-a489-9e9fb9c730b5" (UID: "cd0e43f7-cac1-4f80-a489-9e9fb9c730b5"). InnerVolumeSpecName "kube-api-access-pw8gj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.005180 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd0e43f7-cac1-4f80-a489-9e9fb9c730b5-config-data" (OuterVolumeSpecName: "config-data") pod "cd0e43f7-cac1-4f80-a489-9e9fb9c730b5" (UID: "cd0e43f7-cac1-4f80-a489-9e9fb9c730b5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.006345 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd0e43f7-cac1-4f80-a489-9e9fb9c730b5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd0e43f7-cac1-4f80-a489-9e9fb9c730b5" (UID: "cd0e43f7-cac1-4f80-a489-9e9fb9c730b5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.079045 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd0e43f7-cac1-4f80-a489-9e9fb9c730b5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.079407 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd0e43f7-cac1-4f80-a489-9e9fb9c730b5-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.079421 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pw8gj\" (UniqueName: \"kubernetes.io/projected/cd0e43f7-cac1-4f80-a489-9e9fb9c730b5-kube-api-access-pw8gj\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.407591 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.455514 4959 generic.go:334] "Generic (PLEG): container finished" podID="1fbef70d-f88f-4a9b-bbfd-cabbba6ab107" containerID="6704b54501cf7fad38d8058c3cab86b4f30bf644b305b76a93a808c9c796dfcb" exitCode=0 Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.455558 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1fbef70d-f88f-4a9b-bbfd-cabbba6ab107","Type":"ContainerDied","Data":"6704b54501cf7fad38d8058c3cab86b4f30bf644b305b76a93a808c9c796dfcb"} Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.456886 4959 generic.go:334] "Generic (PLEG): container finished" podID="cd0e43f7-cac1-4f80-a489-9e9fb9c730b5" containerID="18031b1a9be033a4e4fadbf95c756c4a2fa3c53951a4833acd26f02e2ea91e10" exitCode=0 Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.456909 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cd0e43f7-cac1-4f80-a489-9e9fb9c730b5","Type":"ContainerDied","Data":"18031b1a9be033a4e4fadbf95c756c4a2fa3c53951a4833acd26f02e2ea91e10"} Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.456951 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cd0e43f7-cac1-4f80-a489-9e9fb9c730b5","Type":"ContainerDied","Data":"1b09cae32c47905184c09f1d54a92634f54f6e7d5fb7e340f8518531a54f689d"} Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.456952 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.456966 4959 scope.go:117] "RemoveContainer" containerID="18031b1a9be033a4e4fadbf95c756c4a2fa3c53951a4833acd26f02e2ea91e10" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.459967 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"850dd238-e869-4484-9d52-3c67dc241120","Type":"ContainerStarted","Data":"61a0dc6e947b3fe55b6a5a79761e2d991d2cb78a7859501fb1f9fe5e42314a1e"} Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.460596 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.481026 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.493572 4959 scope.go:117] "RemoveContainer" containerID="18031b1a9be033a4e4fadbf95c756c4a2fa3c53951a4833acd26f02e2ea91e10" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.494311 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.494284213 podStartE2EDuration="2.494284213s" podCreationTimestamp="2025-10-07 14:05:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:05:36.483538913 +0000 UTC m=+1218.566943238" watchObservedRunningTime="2025-10-07 14:05:36.494284213 +0000 UTC m=+1218.577688538" Oct 07 14:05:36 crc kubenswrapper[4959]: E1007 14:05:36.496349 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18031b1a9be033a4e4fadbf95c756c4a2fa3c53951a4833acd26f02e2ea91e10\": container with ID starting with 18031b1a9be033a4e4fadbf95c756c4a2fa3c53951a4833acd26f02e2ea91e10 not found: ID does not exist" containerID="18031b1a9be033a4e4fadbf95c756c4a2fa3c53951a4833acd26f02e2ea91e10" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.496391 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18031b1a9be033a4e4fadbf95c756c4a2fa3c53951a4833acd26f02e2ea91e10"} err="failed to get container status \"18031b1a9be033a4e4fadbf95c756c4a2fa3c53951a4833acd26f02e2ea91e10\": rpc error: code = NotFound desc = could not find container \"18031b1a9be033a4e4fadbf95c756c4a2fa3c53951a4833acd26f02e2ea91e10\": container with ID starting with 18031b1a9be033a4e4fadbf95c756c4a2fa3c53951a4833acd26f02e2ea91e10 not found: ID does not exist" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.537162 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.545613 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.553324 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 14:05:36 crc kubenswrapper[4959]: E1007 14:05:36.553818 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fbef70d-f88f-4a9b-bbfd-cabbba6ab107" containerName="nova-api-api" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.553843 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fbef70d-f88f-4a9b-bbfd-cabbba6ab107" containerName="nova-api-api" Oct 07 14:05:36 crc kubenswrapper[4959]: E1007 14:05:36.553873 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd0e43f7-cac1-4f80-a489-9e9fb9c730b5" containerName="nova-scheduler-scheduler" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.553900 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd0e43f7-cac1-4f80-a489-9e9fb9c730b5" containerName="nova-scheduler-scheduler" Oct 07 14:05:36 crc kubenswrapper[4959]: E1007 14:05:36.553919 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fbef70d-f88f-4a9b-bbfd-cabbba6ab107" containerName="nova-api-log" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.553931 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fbef70d-f88f-4a9b-bbfd-cabbba6ab107" containerName="nova-api-log" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.554151 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fbef70d-f88f-4a9b-bbfd-cabbba6ab107" containerName="nova-api-log" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.554174 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fbef70d-f88f-4a9b-bbfd-cabbba6ab107" containerName="nova-api-api" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.554194 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd0e43f7-cac1-4f80-a489-9e9fb9c730b5" containerName="nova-scheduler-scheduler" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.554924 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.557319 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.561198 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.588689 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1fbef70d-f88f-4a9b-bbfd-cabbba6ab107-logs\") pod \"1fbef70d-f88f-4a9b-bbfd-cabbba6ab107\" (UID: \"1fbef70d-f88f-4a9b-bbfd-cabbba6ab107\") " Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.588775 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9jvt\" (UniqueName: \"kubernetes.io/projected/1fbef70d-f88f-4a9b-bbfd-cabbba6ab107-kube-api-access-l9jvt\") pod \"1fbef70d-f88f-4a9b-bbfd-cabbba6ab107\" (UID: \"1fbef70d-f88f-4a9b-bbfd-cabbba6ab107\") " Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.588820 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fbef70d-f88f-4a9b-bbfd-cabbba6ab107-config-data\") pod \"1fbef70d-f88f-4a9b-bbfd-cabbba6ab107\" (UID: \"1fbef70d-f88f-4a9b-bbfd-cabbba6ab107\") " Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.588939 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fbef70d-f88f-4a9b-bbfd-cabbba6ab107-combined-ca-bundle\") pod \"1fbef70d-f88f-4a9b-bbfd-cabbba6ab107\" (UID: \"1fbef70d-f88f-4a9b-bbfd-cabbba6ab107\") " Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.591161 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fbef70d-f88f-4a9b-bbfd-cabbba6ab107-logs" (OuterVolumeSpecName: "logs") pod "1fbef70d-f88f-4a9b-bbfd-cabbba6ab107" (UID: "1fbef70d-f88f-4a9b-bbfd-cabbba6ab107"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.592887 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fbef70d-f88f-4a9b-bbfd-cabbba6ab107-kube-api-access-l9jvt" (OuterVolumeSpecName: "kube-api-access-l9jvt") pod "1fbef70d-f88f-4a9b-bbfd-cabbba6ab107" (UID: "1fbef70d-f88f-4a9b-bbfd-cabbba6ab107"). InnerVolumeSpecName "kube-api-access-l9jvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.619279 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fbef70d-f88f-4a9b-bbfd-cabbba6ab107-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1fbef70d-f88f-4a9b-bbfd-cabbba6ab107" (UID: "1fbef70d-f88f-4a9b-bbfd-cabbba6ab107"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.622702 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fbef70d-f88f-4a9b-bbfd-cabbba6ab107-config-data" (OuterVolumeSpecName: "config-data") pod "1fbef70d-f88f-4a9b-bbfd-cabbba6ab107" (UID: "1fbef70d-f88f-4a9b-bbfd-cabbba6ab107"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.667494 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd0e43f7-cac1-4f80-a489-9e9fb9c730b5" path="/var/lib/kubelet/pods/cd0e43f7-cac1-4f80-a489-9e9fb9c730b5/volumes" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.690927 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14f0652c-3a52-460c-9ebc-86eeba72e77a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"14f0652c-3a52-460c-9ebc-86eeba72e77a\") " pod="openstack/nova-scheduler-0" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.690987 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14f0652c-3a52-460c-9ebc-86eeba72e77a-config-data\") pod \"nova-scheduler-0\" (UID: \"14f0652c-3a52-460c-9ebc-86eeba72e77a\") " pod="openstack/nova-scheduler-0" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.691012 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvls2\" (UniqueName: \"kubernetes.io/projected/14f0652c-3a52-460c-9ebc-86eeba72e77a-kube-api-access-kvls2\") pod \"nova-scheduler-0\" (UID: \"14f0652c-3a52-460c-9ebc-86eeba72e77a\") " pod="openstack/nova-scheduler-0" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.691116 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1fbef70d-f88f-4a9b-bbfd-cabbba6ab107-logs\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.691134 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9jvt\" (UniqueName: \"kubernetes.io/projected/1fbef70d-f88f-4a9b-bbfd-cabbba6ab107-kube-api-access-l9jvt\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.691148 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fbef70d-f88f-4a9b-bbfd-cabbba6ab107-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.691159 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fbef70d-f88f-4a9b-bbfd-cabbba6ab107-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.793038 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14f0652c-3a52-460c-9ebc-86eeba72e77a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"14f0652c-3a52-460c-9ebc-86eeba72e77a\") " pod="openstack/nova-scheduler-0" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.793412 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14f0652c-3a52-460c-9ebc-86eeba72e77a-config-data\") pod \"nova-scheduler-0\" (UID: \"14f0652c-3a52-460c-9ebc-86eeba72e77a\") " pod="openstack/nova-scheduler-0" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.793443 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvls2\" (UniqueName: \"kubernetes.io/projected/14f0652c-3a52-460c-9ebc-86eeba72e77a-kube-api-access-kvls2\") pod \"nova-scheduler-0\" (UID: \"14f0652c-3a52-460c-9ebc-86eeba72e77a\") " pod="openstack/nova-scheduler-0" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.797616 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14f0652c-3a52-460c-9ebc-86eeba72e77a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"14f0652c-3a52-460c-9ebc-86eeba72e77a\") " pod="openstack/nova-scheduler-0" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.797670 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14f0652c-3a52-460c-9ebc-86eeba72e77a-config-data\") pod \"nova-scheduler-0\" (UID: \"14f0652c-3a52-460c-9ebc-86eeba72e77a\") " pod="openstack/nova-scheduler-0" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.810747 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvls2\" (UniqueName: \"kubernetes.io/projected/14f0652c-3a52-460c-9ebc-86eeba72e77a-kube-api-access-kvls2\") pod \"nova-scheduler-0\" (UID: \"14f0652c-3a52-460c-9ebc-86eeba72e77a\") " pod="openstack/nova-scheduler-0" Oct 07 14:05:36 crc kubenswrapper[4959]: I1007 14:05:36.875182 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 14:05:37 crc kubenswrapper[4959]: I1007 14:05:37.341695 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 14:05:37 crc kubenswrapper[4959]: W1007 14:05:37.353514 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod14f0652c_3a52_460c_9ebc_86eeba72e77a.slice/crio-cd76603308801e42635139edd46ec5c0b186132797e4230ec2650b1e3d588c22 WatchSource:0}: Error finding container cd76603308801e42635139edd46ec5c0b186132797e4230ec2650b1e3d588c22: Status 404 returned error can't find the container with id cd76603308801e42635139edd46ec5c0b186132797e4230ec2650b1e3d588c22 Oct 07 14:05:37 crc kubenswrapper[4959]: I1007 14:05:37.468393 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"14f0652c-3a52-460c-9ebc-86eeba72e77a","Type":"ContainerStarted","Data":"cd76603308801e42635139edd46ec5c0b186132797e4230ec2650b1e3d588c22"} Oct 07 14:05:37 crc kubenswrapper[4959]: I1007 14:05:37.470954 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1fbef70d-f88f-4a9b-bbfd-cabbba6ab107","Type":"ContainerDied","Data":"6f4f0e8d7f0d2b1b6316cb45476dc9fd3fc8288db131e12a4f641b184f8eae11"} Oct 07 14:05:37 crc kubenswrapper[4959]: I1007 14:05:37.470991 4959 scope.go:117] "RemoveContainer" containerID="6704b54501cf7fad38d8058c3cab86b4f30bf644b305b76a93a808c9c796dfcb" Oct 07 14:05:37 crc kubenswrapper[4959]: I1007 14:05:37.471088 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 14:05:37 crc kubenswrapper[4959]: I1007 14:05:37.490115 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 07 14:05:37 crc kubenswrapper[4959]: I1007 14:05:37.497228 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 07 14:05:37 crc kubenswrapper[4959]: I1007 14:05:37.511517 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 07 14:05:37 crc kubenswrapper[4959]: I1007 14:05:37.512987 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 14:05:37 crc kubenswrapper[4959]: I1007 14:05:37.520733 4959 scope.go:117] "RemoveContainer" containerID="c72e1ceefde275c1043f6aa2ce2b4e66addae8eb914dad971b7ade250648dff1" Oct 07 14:05:37 crc kubenswrapper[4959]: I1007 14:05:37.520994 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 07 14:05:37 crc kubenswrapper[4959]: I1007 14:05:37.530307 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 14:05:37 crc kubenswrapper[4959]: I1007 14:05:37.709663 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8ce7f2f-cbb7-4daa-9955-d71c900e32d5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a8ce7f2f-cbb7-4daa-9955-d71c900e32d5\") " pod="openstack/nova-api-0" Oct 07 14:05:37 crc kubenswrapper[4959]: I1007 14:05:37.710093 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8ce7f2f-cbb7-4daa-9955-d71c900e32d5-logs\") pod \"nova-api-0\" (UID: \"a8ce7f2f-cbb7-4daa-9955-d71c900e32d5\") " pod="openstack/nova-api-0" Oct 07 14:05:37 crc kubenswrapper[4959]: I1007 14:05:37.710330 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8ce7f2f-cbb7-4daa-9955-d71c900e32d5-config-data\") pod \"nova-api-0\" (UID: \"a8ce7f2f-cbb7-4daa-9955-d71c900e32d5\") " pod="openstack/nova-api-0" Oct 07 14:05:37 crc kubenswrapper[4959]: I1007 14:05:37.710449 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdvpz\" (UniqueName: \"kubernetes.io/projected/a8ce7f2f-cbb7-4daa-9955-d71c900e32d5-kube-api-access-zdvpz\") pod \"nova-api-0\" (UID: \"a8ce7f2f-cbb7-4daa-9955-d71c900e32d5\") " pod="openstack/nova-api-0" Oct 07 14:05:37 crc kubenswrapper[4959]: I1007 14:05:37.808662 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 07 14:05:37 crc kubenswrapper[4959]: I1007 14:05:37.808702 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 07 14:05:37 crc kubenswrapper[4959]: I1007 14:05:37.811288 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8ce7f2f-cbb7-4daa-9955-d71c900e32d5-config-data\") pod \"nova-api-0\" (UID: \"a8ce7f2f-cbb7-4daa-9955-d71c900e32d5\") " pod="openstack/nova-api-0" Oct 07 14:05:37 crc kubenswrapper[4959]: I1007 14:05:37.811322 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdvpz\" (UniqueName: \"kubernetes.io/projected/a8ce7f2f-cbb7-4daa-9955-d71c900e32d5-kube-api-access-zdvpz\") pod \"nova-api-0\" (UID: \"a8ce7f2f-cbb7-4daa-9955-d71c900e32d5\") " pod="openstack/nova-api-0" Oct 07 14:05:37 crc kubenswrapper[4959]: I1007 14:05:37.811543 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8ce7f2f-cbb7-4daa-9955-d71c900e32d5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a8ce7f2f-cbb7-4daa-9955-d71c900e32d5\") " pod="openstack/nova-api-0" Oct 07 14:05:37 crc kubenswrapper[4959]: I1007 14:05:37.811597 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8ce7f2f-cbb7-4daa-9955-d71c900e32d5-logs\") pod \"nova-api-0\" (UID: \"a8ce7f2f-cbb7-4daa-9955-d71c900e32d5\") " pod="openstack/nova-api-0" Oct 07 14:05:37 crc kubenswrapper[4959]: I1007 14:05:37.812114 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8ce7f2f-cbb7-4daa-9955-d71c900e32d5-logs\") pod \"nova-api-0\" (UID: \"a8ce7f2f-cbb7-4daa-9955-d71c900e32d5\") " pod="openstack/nova-api-0" Oct 07 14:05:37 crc kubenswrapper[4959]: I1007 14:05:37.819953 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8ce7f2f-cbb7-4daa-9955-d71c900e32d5-config-data\") pod \"nova-api-0\" (UID: \"a8ce7f2f-cbb7-4daa-9955-d71c900e32d5\") " pod="openstack/nova-api-0" Oct 07 14:05:37 crc kubenswrapper[4959]: I1007 14:05:37.821759 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8ce7f2f-cbb7-4daa-9955-d71c900e32d5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a8ce7f2f-cbb7-4daa-9955-d71c900e32d5\") " pod="openstack/nova-api-0" Oct 07 14:05:37 crc kubenswrapper[4959]: I1007 14:05:37.832083 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdvpz\" (UniqueName: \"kubernetes.io/projected/a8ce7f2f-cbb7-4daa-9955-d71c900e32d5-kube-api-access-zdvpz\") pod \"nova-api-0\" (UID: \"a8ce7f2f-cbb7-4daa-9955-d71c900e32d5\") " pod="openstack/nova-api-0" Oct 07 14:05:37 crc kubenswrapper[4959]: I1007 14:05:37.835611 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 14:05:38 crc kubenswrapper[4959]: I1007 14:05:38.292863 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 14:05:38 crc kubenswrapper[4959]: W1007 14:05:38.300660 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda8ce7f2f_cbb7_4daa_9955_d71c900e32d5.slice/crio-44acfb12438fe547c9f197f7df293407a059f6d424e32f0609b7673c5910b433 WatchSource:0}: Error finding container 44acfb12438fe547c9f197f7df293407a059f6d424e32f0609b7673c5910b433: Status 404 returned error can't find the container with id 44acfb12438fe547c9f197f7df293407a059f6d424e32f0609b7673c5910b433 Oct 07 14:05:38 crc kubenswrapper[4959]: I1007 14:05:38.484397 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"14f0652c-3a52-460c-9ebc-86eeba72e77a","Type":"ContainerStarted","Data":"48cf7552cac7401d7fb488ed8ed9ab98b4408a52de7d47873248702d94af9ad5"} Oct 07 14:05:38 crc kubenswrapper[4959]: I1007 14:05:38.489035 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a8ce7f2f-cbb7-4daa-9955-d71c900e32d5","Type":"ContainerStarted","Data":"0ff620573f010a5a7e6cd46a4ed873136ab51b6ecaaf0a166bece48e3f98c4a5"} Oct 07 14:05:38 crc kubenswrapper[4959]: I1007 14:05:38.489371 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a8ce7f2f-cbb7-4daa-9955-d71c900e32d5","Type":"ContainerStarted","Data":"44acfb12438fe547c9f197f7df293407a059f6d424e32f0609b7673c5910b433"} Oct 07 14:05:38 crc kubenswrapper[4959]: I1007 14:05:38.516875 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.516848113 podStartE2EDuration="2.516848113s" podCreationTimestamp="2025-10-07 14:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:05:38.507638542 +0000 UTC m=+1220.591042867" watchObservedRunningTime="2025-10-07 14:05:38.516848113 +0000 UTC m=+1220.600252468" Oct 07 14:05:38 crc kubenswrapper[4959]: I1007 14:05:38.679838 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fbef70d-f88f-4a9b-bbfd-cabbba6ab107" path="/var/lib/kubelet/pods/1fbef70d-f88f-4a9b-bbfd-cabbba6ab107/volumes" Oct 07 14:05:39 crc kubenswrapper[4959]: I1007 14:05:39.501659 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a8ce7f2f-cbb7-4daa-9955-d71c900e32d5","Type":"ContainerStarted","Data":"faa17a2a452375a503f4d9e6af1a500f6c57913d483377e6df70906306f2892d"} Oct 07 14:05:39 crc kubenswrapper[4959]: I1007 14:05:39.530339 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.530323093 podStartE2EDuration="2.530323093s" podCreationTimestamp="2025-10-07 14:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:05:39.523210155 +0000 UTC m=+1221.606614520" watchObservedRunningTime="2025-10-07 14:05:39.530323093 +0000 UTC m=+1221.613727418" Oct 07 14:05:41 crc kubenswrapper[4959]: I1007 14:05:41.875410 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 07 14:05:42 crc kubenswrapper[4959]: I1007 14:05:42.809007 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 07 14:05:42 crc kubenswrapper[4959]: I1007 14:05:42.809077 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 07 14:05:43 crc kubenswrapper[4959]: I1007 14:05:43.824270 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="0a9c0a78-b2ae-4edb-8b49-be96b1b0731f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.184:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 07 14:05:43 crc kubenswrapper[4959]: I1007 14:05:43.824876 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="0a9c0a78-b2ae-4edb-8b49-be96b1b0731f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.184:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 07 14:05:44 crc kubenswrapper[4959]: I1007 14:05:44.889214 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 07 14:05:46 crc kubenswrapper[4959]: I1007 14:05:46.876552 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 07 14:05:46 crc kubenswrapper[4959]: I1007 14:05:46.922849 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 07 14:05:47 crc kubenswrapper[4959]: I1007 14:05:47.610618 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 07 14:05:47 crc kubenswrapper[4959]: I1007 14:05:47.836373 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 07 14:05:47 crc kubenswrapper[4959]: I1007 14:05:47.836908 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 07 14:05:48 crc kubenswrapper[4959]: I1007 14:05:48.920374 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a8ce7f2f-cbb7-4daa-9955-d71c900e32d5" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.187:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 14:05:48 crc kubenswrapper[4959]: I1007 14:05:48.920391 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a8ce7f2f-cbb7-4daa-9955-d71c900e32d5" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.187:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 07 14:05:52 crc kubenswrapper[4959]: I1007 14:05:52.815757 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 07 14:05:52 crc kubenswrapper[4959]: I1007 14:05:52.819600 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 07 14:05:52 crc kubenswrapper[4959]: I1007 14:05:52.826673 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 07 14:05:53 crc kubenswrapper[4959]: I1007 14:05:53.650014 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 07 14:05:54 crc kubenswrapper[4959]: I1007 14:05:54.598040 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:54 crc kubenswrapper[4959]: I1007 14:05:54.655887 4959 generic.go:334] "Generic (PLEG): container finished" podID="353cb13b-1538-496c-9a26-47f90b784aa4" containerID="b2f9ae43c476cc206537b7de7cdda2fad5b2b7a01f98e0d9c4f2ec7e29936e54" exitCode=137 Oct 07 14:05:54 crc kubenswrapper[4959]: I1007 14:05:54.656247 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:54 crc kubenswrapper[4959]: I1007 14:05:54.676049 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"353cb13b-1538-496c-9a26-47f90b784aa4","Type":"ContainerDied","Data":"b2f9ae43c476cc206537b7de7cdda2fad5b2b7a01f98e0d9c4f2ec7e29936e54"} Oct 07 14:05:54 crc kubenswrapper[4959]: I1007 14:05:54.676184 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"353cb13b-1538-496c-9a26-47f90b784aa4","Type":"ContainerDied","Data":"8138d5710fb875595ba16f5ed50ccd91512a4e403ad3f7fba85c2eb2a4f71379"} Oct 07 14:05:54 crc kubenswrapper[4959]: I1007 14:05:54.676225 4959 scope.go:117] "RemoveContainer" containerID="b2f9ae43c476cc206537b7de7cdda2fad5b2b7a01f98e0d9c4f2ec7e29936e54" Oct 07 14:05:54 crc kubenswrapper[4959]: I1007 14:05:54.710529 4959 scope.go:117] "RemoveContainer" containerID="b2f9ae43c476cc206537b7de7cdda2fad5b2b7a01f98e0d9c4f2ec7e29936e54" Oct 07 14:05:54 crc kubenswrapper[4959]: E1007 14:05:54.711073 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2f9ae43c476cc206537b7de7cdda2fad5b2b7a01f98e0d9c4f2ec7e29936e54\": container with ID starting with b2f9ae43c476cc206537b7de7cdda2fad5b2b7a01f98e0d9c4f2ec7e29936e54 not found: ID does not exist" containerID="b2f9ae43c476cc206537b7de7cdda2fad5b2b7a01f98e0d9c4f2ec7e29936e54" Oct 07 14:05:54 crc kubenswrapper[4959]: I1007 14:05:54.711158 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2f9ae43c476cc206537b7de7cdda2fad5b2b7a01f98e0d9c4f2ec7e29936e54"} err="failed to get container status \"b2f9ae43c476cc206537b7de7cdda2fad5b2b7a01f98e0d9c4f2ec7e29936e54\": rpc error: code = NotFound desc = could not find container \"b2f9ae43c476cc206537b7de7cdda2fad5b2b7a01f98e0d9c4f2ec7e29936e54\": container with ID starting with b2f9ae43c476cc206537b7de7cdda2fad5b2b7a01f98e0d9c4f2ec7e29936e54 not found: ID does not exist" Oct 07 14:05:54 crc kubenswrapper[4959]: I1007 14:05:54.720808 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/353cb13b-1538-496c-9a26-47f90b784aa4-combined-ca-bundle\") pod \"353cb13b-1538-496c-9a26-47f90b784aa4\" (UID: \"353cb13b-1538-496c-9a26-47f90b784aa4\") " Oct 07 14:05:54 crc kubenswrapper[4959]: I1007 14:05:54.720906 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/353cb13b-1538-496c-9a26-47f90b784aa4-config-data\") pod \"353cb13b-1538-496c-9a26-47f90b784aa4\" (UID: \"353cb13b-1538-496c-9a26-47f90b784aa4\") " Oct 07 14:05:54 crc kubenswrapper[4959]: I1007 14:05:54.721053 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-678qv\" (UniqueName: \"kubernetes.io/projected/353cb13b-1538-496c-9a26-47f90b784aa4-kube-api-access-678qv\") pod \"353cb13b-1538-496c-9a26-47f90b784aa4\" (UID: \"353cb13b-1538-496c-9a26-47f90b784aa4\") " Oct 07 14:05:54 crc kubenswrapper[4959]: I1007 14:05:54.728707 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/353cb13b-1538-496c-9a26-47f90b784aa4-kube-api-access-678qv" (OuterVolumeSpecName: "kube-api-access-678qv") pod "353cb13b-1538-496c-9a26-47f90b784aa4" (UID: "353cb13b-1538-496c-9a26-47f90b784aa4"). InnerVolumeSpecName "kube-api-access-678qv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:05:54 crc kubenswrapper[4959]: I1007 14:05:54.760507 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/353cb13b-1538-496c-9a26-47f90b784aa4-config-data" (OuterVolumeSpecName: "config-data") pod "353cb13b-1538-496c-9a26-47f90b784aa4" (UID: "353cb13b-1538-496c-9a26-47f90b784aa4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:05:54 crc kubenswrapper[4959]: I1007 14:05:54.762809 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/353cb13b-1538-496c-9a26-47f90b784aa4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "353cb13b-1538-496c-9a26-47f90b784aa4" (UID: "353cb13b-1538-496c-9a26-47f90b784aa4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:05:54 crc kubenswrapper[4959]: I1007 14:05:54.823889 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-678qv\" (UniqueName: \"kubernetes.io/projected/353cb13b-1538-496c-9a26-47f90b784aa4-kube-api-access-678qv\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:54 crc kubenswrapper[4959]: I1007 14:05:54.823928 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/353cb13b-1538-496c-9a26-47f90b784aa4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:54 crc kubenswrapper[4959]: I1007 14:05:54.823937 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/353cb13b-1538-496c-9a26-47f90b784aa4-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:05:54 crc kubenswrapper[4959]: I1007 14:05:54.990928 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 14:05:55 crc kubenswrapper[4959]: I1007 14:05:55.001553 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 14:05:55 crc kubenswrapper[4959]: I1007 14:05:55.016819 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 14:05:55 crc kubenswrapper[4959]: E1007 14:05:55.018807 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="353cb13b-1538-496c-9a26-47f90b784aa4" containerName="nova-cell1-novncproxy-novncproxy" Oct 07 14:05:55 crc kubenswrapper[4959]: I1007 14:05:55.018842 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="353cb13b-1538-496c-9a26-47f90b784aa4" containerName="nova-cell1-novncproxy-novncproxy" Oct 07 14:05:55 crc kubenswrapper[4959]: I1007 14:05:55.019764 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="353cb13b-1538-496c-9a26-47f90b784aa4" containerName="nova-cell1-novncproxy-novncproxy" Oct 07 14:05:55 crc kubenswrapper[4959]: I1007 14:05:55.020904 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:55 crc kubenswrapper[4959]: I1007 14:05:55.031928 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88e679d5-b7cb-4e49-9f01-ab4ceacd43f1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"88e679d5-b7cb-4e49-9f01-ab4ceacd43f1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:55 crc kubenswrapper[4959]: I1007 14:05:55.032052 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/88e679d5-b7cb-4e49-9f01-ab4ceacd43f1-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"88e679d5-b7cb-4e49-9f01-ab4ceacd43f1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:55 crc kubenswrapper[4959]: I1007 14:05:55.032211 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88e679d5-b7cb-4e49-9f01-ab4ceacd43f1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"88e679d5-b7cb-4e49-9f01-ab4ceacd43f1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:55 crc kubenswrapper[4959]: I1007 14:05:55.032304 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qq27\" (UniqueName: \"kubernetes.io/projected/88e679d5-b7cb-4e49-9f01-ab4ceacd43f1-kube-api-access-4qq27\") pod \"nova-cell1-novncproxy-0\" (UID: \"88e679d5-b7cb-4e49-9f01-ab4ceacd43f1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:55 crc kubenswrapper[4959]: I1007 14:05:55.032423 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/88e679d5-b7cb-4e49-9f01-ab4ceacd43f1-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"88e679d5-b7cb-4e49-9f01-ab4ceacd43f1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:55 crc kubenswrapper[4959]: I1007 14:05:55.036577 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 07 14:05:55 crc kubenswrapper[4959]: I1007 14:05:55.036828 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 07 14:05:55 crc kubenswrapper[4959]: I1007 14:05:55.036864 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 07 14:05:55 crc kubenswrapper[4959]: I1007 14:05:55.039137 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 14:05:55 crc kubenswrapper[4959]: I1007 14:05:55.134370 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88e679d5-b7cb-4e49-9f01-ab4ceacd43f1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"88e679d5-b7cb-4e49-9f01-ab4ceacd43f1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:55 crc kubenswrapper[4959]: I1007 14:05:55.134535 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qq27\" (UniqueName: \"kubernetes.io/projected/88e679d5-b7cb-4e49-9f01-ab4ceacd43f1-kube-api-access-4qq27\") pod \"nova-cell1-novncproxy-0\" (UID: \"88e679d5-b7cb-4e49-9f01-ab4ceacd43f1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:55 crc kubenswrapper[4959]: I1007 14:05:55.134620 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/88e679d5-b7cb-4e49-9f01-ab4ceacd43f1-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"88e679d5-b7cb-4e49-9f01-ab4ceacd43f1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:55 crc kubenswrapper[4959]: I1007 14:05:55.134739 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88e679d5-b7cb-4e49-9f01-ab4ceacd43f1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"88e679d5-b7cb-4e49-9f01-ab4ceacd43f1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:55 crc kubenswrapper[4959]: I1007 14:05:55.134837 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/88e679d5-b7cb-4e49-9f01-ab4ceacd43f1-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"88e679d5-b7cb-4e49-9f01-ab4ceacd43f1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:55 crc kubenswrapper[4959]: I1007 14:05:55.138517 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/88e679d5-b7cb-4e49-9f01-ab4ceacd43f1-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"88e679d5-b7cb-4e49-9f01-ab4ceacd43f1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:55 crc kubenswrapper[4959]: I1007 14:05:55.145548 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88e679d5-b7cb-4e49-9f01-ab4ceacd43f1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"88e679d5-b7cb-4e49-9f01-ab4ceacd43f1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:55 crc kubenswrapper[4959]: I1007 14:05:55.145864 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88e679d5-b7cb-4e49-9f01-ab4ceacd43f1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"88e679d5-b7cb-4e49-9f01-ab4ceacd43f1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:55 crc kubenswrapper[4959]: I1007 14:05:55.146569 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/88e679d5-b7cb-4e49-9f01-ab4ceacd43f1-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"88e679d5-b7cb-4e49-9f01-ab4ceacd43f1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:55 crc kubenswrapper[4959]: I1007 14:05:55.150475 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qq27\" (UniqueName: \"kubernetes.io/projected/88e679d5-b7cb-4e49-9f01-ab4ceacd43f1-kube-api-access-4qq27\") pod \"nova-cell1-novncproxy-0\" (UID: \"88e679d5-b7cb-4e49-9f01-ab4ceacd43f1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:55 crc kubenswrapper[4959]: I1007 14:05:55.351706 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:05:55 crc kubenswrapper[4959]: I1007 14:05:55.770845 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 07 14:05:56 crc kubenswrapper[4959]: I1007 14:05:56.673716 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="353cb13b-1538-496c-9a26-47f90b784aa4" path="/var/lib/kubelet/pods/353cb13b-1538-496c-9a26-47f90b784aa4/volumes" Oct 07 14:05:56 crc kubenswrapper[4959]: I1007 14:05:56.681083 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"88e679d5-b7cb-4e49-9f01-ab4ceacd43f1","Type":"ContainerStarted","Data":"746d887ff499c6a164278ccf8ef5659a997b774eeac634b2974412372478491b"} Oct 07 14:05:56 crc kubenswrapper[4959]: I1007 14:05:56.681167 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"88e679d5-b7cb-4e49-9f01-ab4ceacd43f1","Type":"ContainerStarted","Data":"57c9d42869f959aed6b9841d42e1e4c63a9565524ef9d3b8d6f4c8b889234057"} Oct 07 14:05:56 crc kubenswrapper[4959]: I1007 14:05:56.715205 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.715183569 podStartE2EDuration="2.715183569s" podCreationTimestamp="2025-10-07 14:05:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:05:56.701115526 +0000 UTC m=+1238.784519851" watchObservedRunningTime="2025-10-07 14:05:56.715183569 +0000 UTC m=+1238.798587894" Oct 07 14:05:57 crc kubenswrapper[4959]: I1007 14:05:57.840394 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 07 14:05:57 crc kubenswrapper[4959]: I1007 14:05:57.841751 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 07 14:05:57 crc kubenswrapper[4959]: I1007 14:05:57.844277 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 07 14:05:57 crc kubenswrapper[4959]: I1007 14:05:57.844670 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 07 14:05:58 crc kubenswrapper[4959]: I1007 14:05:58.701334 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 07 14:05:58 crc kubenswrapper[4959]: I1007 14:05:58.706280 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 07 14:05:58 crc kubenswrapper[4959]: I1007 14:05:58.888554 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-665946c669-2wpls"] Oct 07 14:05:58 crc kubenswrapper[4959]: I1007 14:05:58.891019 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-665946c669-2wpls" Oct 07 14:05:58 crc kubenswrapper[4959]: I1007 14:05:58.917403 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-665946c669-2wpls"] Oct 07 14:05:59 crc kubenswrapper[4959]: I1007 14:05:59.022630 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-ovsdbserver-nb\") pod \"dnsmasq-dns-665946c669-2wpls\" (UID: \"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a\") " pod="openstack/dnsmasq-dns-665946c669-2wpls" Oct 07 14:05:59 crc kubenswrapper[4959]: I1007 14:05:59.022691 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-ovsdbserver-sb\") pod \"dnsmasq-dns-665946c669-2wpls\" (UID: \"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a\") " pod="openstack/dnsmasq-dns-665946c669-2wpls" Oct 07 14:05:59 crc kubenswrapper[4959]: I1007 14:05:59.022774 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-config\") pod \"dnsmasq-dns-665946c669-2wpls\" (UID: \"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a\") " pod="openstack/dnsmasq-dns-665946c669-2wpls" Oct 07 14:05:59 crc kubenswrapper[4959]: I1007 14:05:59.022890 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nd6l\" (UniqueName: \"kubernetes.io/projected/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-kube-api-access-6nd6l\") pod \"dnsmasq-dns-665946c669-2wpls\" (UID: \"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a\") " pod="openstack/dnsmasq-dns-665946c669-2wpls" Oct 07 14:05:59 crc kubenswrapper[4959]: I1007 14:05:59.022965 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-dns-svc\") pod \"dnsmasq-dns-665946c669-2wpls\" (UID: \"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a\") " pod="openstack/dnsmasq-dns-665946c669-2wpls" Oct 07 14:05:59 crc kubenswrapper[4959]: I1007 14:05:59.124598 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nd6l\" (UniqueName: \"kubernetes.io/projected/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-kube-api-access-6nd6l\") pod \"dnsmasq-dns-665946c669-2wpls\" (UID: \"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a\") " pod="openstack/dnsmasq-dns-665946c669-2wpls" Oct 07 14:05:59 crc kubenswrapper[4959]: I1007 14:05:59.124682 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-dns-svc\") pod \"dnsmasq-dns-665946c669-2wpls\" (UID: \"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a\") " pod="openstack/dnsmasq-dns-665946c669-2wpls" Oct 07 14:05:59 crc kubenswrapper[4959]: I1007 14:05:59.124770 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-ovsdbserver-nb\") pod \"dnsmasq-dns-665946c669-2wpls\" (UID: \"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a\") " pod="openstack/dnsmasq-dns-665946c669-2wpls" Oct 07 14:05:59 crc kubenswrapper[4959]: I1007 14:05:59.124796 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-ovsdbserver-sb\") pod \"dnsmasq-dns-665946c669-2wpls\" (UID: \"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a\") " pod="openstack/dnsmasq-dns-665946c669-2wpls" Oct 07 14:05:59 crc kubenswrapper[4959]: I1007 14:05:59.124825 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-config\") pod \"dnsmasq-dns-665946c669-2wpls\" (UID: \"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a\") " pod="openstack/dnsmasq-dns-665946c669-2wpls" Oct 07 14:05:59 crc kubenswrapper[4959]: I1007 14:05:59.125956 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-dns-svc\") pod \"dnsmasq-dns-665946c669-2wpls\" (UID: \"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a\") " pod="openstack/dnsmasq-dns-665946c669-2wpls" Oct 07 14:05:59 crc kubenswrapper[4959]: I1007 14:05:59.125970 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-ovsdbserver-nb\") pod \"dnsmasq-dns-665946c669-2wpls\" (UID: \"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a\") " pod="openstack/dnsmasq-dns-665946c669-2wpls" Oct 07 14:05:59 crc kubenswrapper[4959]: I1007 14:05:59.125955 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-ovsdbserver-sb\") pod \"dnsmasq-dns-665946c669-2wpls\" (UID: \"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a\") " pod="openstack/dnsmasq-dns-665946c669-2wpls" Oct 07 14:05:59 crc kubenswrapper[4959]: I1007 14:05:59.126467 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-config\") pod \"dnsmasq-dns-665946c669-2wpls\" (UID: \"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a\") " pod="openstack/dnsmasq-dns-665946c669-2wpls" Oct 07 14:05:59 crc kubenswrapper[4959]: I1007 14:05:59.149530 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nd6l\" (UniqueName: \"kubernetes.io/projected/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-kube-api-access-6nd6l\") pod \"dnsmasq-dns-665946c669-2wpls\" (UID: \"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a\") " pod="openstack/dnsmasq-dns-665946c669-2wpls" Oct 07 14:05:59 crc kubenswrapper[4959]: I1007 14:05:59.233337 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-665946c669-2wpls" Oct 07 14:05:59 crc kubenswrapper[4959]: W1007 14:05:59.718952 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda2e9e1ab_2ed3_4fa6_8d20_669cc7088b1a.slice/crio-8cef8f2bd89b8544ebbd654c9e764fc41b79ec608e297d827ba08744df44c220 WatchSource:0}: Error finding container 8cef8f2bd89b8544ebbd654c9e764fc41b79ec608e297d827ba08744df44c220: Status 404 returned error can't find the container with id 8cef8f2bd89b8544ebbd654c9e764fc41b79ec608e297d827ba08744df44c220 Oct 07 14:05:59 crc kubenswrapper[4959]: I1007 14:05:59.719297 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-665946c669-2wpls"] Oct 07 14:06:00 crc kubenswrapper[4959]: I1007 14:06:00.351754 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:06:00 crc kubenswrapper[4959]: I1007 14:06:00.724468 4959 generic.go:334] "Generic (PLEG): container finished" podID="a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a" containerID="effeeb87aac91c6f207ee097e5ddf40065b20990f8649ac951a5bdbd6b4ba804" exitCode=0 Oct 07 14:06:00 crc kubenswrapper[4959]: I1007 14:06:00.725439 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-665946c669-2wpls" event={"ID":"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a","Type":"ContainerDied","Data":"effeeb87aac91c6f207ee097e5ddf40065b20990f8649ac951a5bdbd6b4ba804"} Oct 07 14:06:00 crc kubenswrapper[4959]: I1007 14:06:00.725487 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-665946c669-2wpls" event={"ID":"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a","Type":"ContainerStarted","Data":"8cef8f2bd89b8544ebbd654c9e764fc41b79ec608e297d827ba08744df44c220"} Oct 07 14:06:01 crc kubenswrapper[4959]: I1007 14:06:01.008146 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:06:01 crc kubenswrapper[4959]: I1007 14:06:01.008431 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="80b77f5f-b572-4777-9912-800cc7886ddc" containerName="ceilometer-central-agent" containerID="cri-o://c1cacf38716e17dbbf7bd68462d448bdecbf3780bc758d010f17d82736b5c3ae" gracePeriod=30 Oct 07 14:06:01 crc kubenswrapper[4959]: I1007 14:06:01.008480 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="80b77f5f-b572-4777-9912-800cc7886ddc" containerName="sg-core" containerID="cri-o://394f73fb3149de229dcb4af397e77e4a465d315ea05e727bfa839d6025fcdb05" gracePeriod=30 Oct 07 14:06:01 crc kubenswrapper[4959]: I1007 14:06:01.008518 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="80b77f5f-b572-4777-9912-800cc7886ddc" containerName="proxy-httpd" containerID="cri-o://a9eb1598f0d4f84a7bfc717906de92ea9e5a67b5f8a526d14bc6cb44f27d0100" gracePeriod=30 Oct 07 14:06:01 crc kubenswrapper[4959]: I1007 14:06:01.008507 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="80b77f5f-b572-4777-9912-800cc7886ddc" containerName="ceilometer-notification-agent" containerID="cri-o://d083fcb14061bd5b6d3657744d575379b4a02f239d3e5703ba088b08b6136427" gracePeriod=30 Oct 07 14:06:01 crc kubenswrapper[4959]: I1007 14:06:01.088944 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 07 14:06:01 crc kubenswrapper[4959]: I1007 14:06:01.737658 4959 generic.go:334] "Generic (PLEG): container finished" podID="80b77f5f-b572-4777-9912-800cc7886ddc" containerID="a9eb1598f0d4f84a7bfc717906de92ea9e5a67b5f8a526d14bc6cb44f27d0100" exitCode=0 Oct 07 14:06:01 crc kubenswrapper[4959]: I1007 14:06:01.737692 4959 generic.go:334] "Generic (PLEG): container finished" podID="80b77f5f-b572-4777-9912-800cc7886ddc" containerID="394f73fb3149de229dcb4af397e77e4a465d315ea05e727bfa839d6025fcdb05" exitCode=2 Oct 07 14:06:01 crc kubenswrapper[4959]: I1007 14:06:01.737701 4959 generic.go:334] "Generic (PLEG): container finished" podID="80b77f5f-b572-4777-9912-800cc7886ddc" containerID="c1cacf38716e17dbbf7bd68462d448bdecbf3780bc758d010f17d82736b5c3ae" exitCode=0 Oct 07 14:06:01 crc kubenswrapper[4959]: I1007 14:06:01.738413 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"80b77f5f-b572-4777-9912-800cc7886ddc","Type":"ContainerDied","Data":"a9eb1598f0d4f84a7bfc717906de92ea9e5a67b5f8a526d14bc6cb44f27d0100"} Oct 07 14:06:01 crc kubenswrapper[4959]: I1007 14:06:01.738542 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"80b77f5f-b572-4777-9912-800cc7886ddc","Type":"ContainerDied","Data":"394f73fb3149de229dcb4af397e77e4a465d315ea05e727bfa839d6025fcdb05"} Oct 07 14:06:01 crc kubenswrapper[4959]: I1007 14:06:01.738558 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"80b77f5f-b572-4777-9912-800cc7886ddc","Type":"ContainerDied","Data":"c1cacf38716e17dbbf7bd68462d448bdecbf3780bc758d010f17d82736b5c3ae"} Oct 07 14:06:01 crc kubenswrapper[4959]: I1007 14:06:01.739134 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-665946c669-2wpls" event={"ID":"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a","Type":"ContainerStarted","Data":"527e7a4af5611fbce86419259a551f2f2efc05cc81b0f1179c3e5c0a1d9dd965"} Oct 07 14:06:01 crc kubenswrapper[4959]: I1007 14:06:01.739174 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a8ce7f2f-cbb7-4daa-9955-d71c900e32d5" containerName="nova-api-log" containerID="cri-o://0ff620573f010a5a7e6cd46a4ed873136ab51b6ecaaf0a166bece48e3f98c4a5" gracePeriod=30 Oct 07 14:06:01 crc kubenswrapper[4959]: I1007 14:06:01.739262 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a8ce7f2f-cbb7-4daa-9955-d71c900e32d5" containerName="nova-api-api" containerID="cri-o://faa17a2a452375a503f4d9e6af1a500f6c57913d483377e6df70906306f2892d" gracePeriod=30 Oct 07 14:06:01 crc kubenswrapper[4959]: I1007 14:06:01.739565 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-665946c669-2wpls" Oct 07 14:06:01 crc kubenswrapper[4959]: I1007 14:06:01.780132 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-665946c669-2wpls" podStartSLOduration=3.78009425 podStartE2EDuration="3.78009425s" podCreationTimestamp="2025-10-07 14:05:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:06:01.774723035 +0000 UTC m=+1243.858127370" watchObservedRunningTime="2025-10-07 14:06:01.78009425 +0000 UTC m=+1243.863498595" Oct 07 14:06:02 crc kubenswrapper[4959]: I1007 14:06:02.752488 4959 generic.go:334] "Generic (PLEG): container finished" podID="80b77f5f-b572-4777-9912-800cc7886ddc" containerID="d083fcb14061bd5b6d3657744d575379b4a02f239d3e5703ba088b08b6136427" exitCode=0 Oct 07 14:06:02 crc kubenswrapper[4959]: I1007 14:06:02.752529 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"80b77f5f-b572-4777-9912-800cc7886ddc","Type":"ContainerDied","Data":"d083fcb14061bd5b6d3657744d575379b4a02f239d3e5703ba088b08b6136427"} Oct 07 14:06:02 crc kubenswrapper[4959]: I1007 14:06:02.754663 4959 generic.go:334] "Generic (PLEG): container finished" podID="a8ce7f2f-cbb7-4daa-9955-d71c900e32d5" containerID="0ff620573f010a5a7e6cd46a4ed873136ab51b6ecaaf0a166bece48e3f98c4a5" exitCode=143 Oct 07 14:06:02 crc kubenswrapper[4959]: I1007 14:06:02.754755 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a8ce7f2f-cbb7-4daa-9955-d71c900e32d5","Type":"ContainerDied","Data":"0ff620573f010a5a7e6cd46a4ed873136ab51b6ecaaf0a166bece48e3f98c4a5"} Oct 07 14:06:02 crc kubenswrapper[4959]: I1007 14:06:02.849112 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.001506 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-scripts\") pod \"80b77f5f-b572-4777-9912-800cc7886ddc\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.002617 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-config-data\") pod \"80b77f5f-b572-4777-9912-800cc7886ddc\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.002675 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwqsr\" (UniqueName: \"kubernetes.io/projected/80b77f5f-b572-4777-9912-800cc7886ddc-kube-api-access-bwqsr\") pod \"80b77f5f-b572-4777-9912-800cc7886ddc\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.002734 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-sg-core-conf-yaml\") pod \"80b77f5f-b572-4777-9912-800cc7886ddc\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.002754 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-ceilometer-tls-certs\") pod \"80b77f5f-b572-4777-9912-800cc7886ddc\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.002776 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80b77f5f-b572-4777-9912-800cc7886ddc-run-httpd\") pod \"80b77f5f-b572-4777-9912-800cc7886ddc\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.002832 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80b77f5f-b572-4777-9912-800cc7886ddc-log-httpd\") pod \"80b77f5f-b572-4777-9912-800cc7886ddc\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.002872 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-combined-ca-bundle\") pod \"80b77f5f-b572-4777-9912-800cc7886ddc\" (UID: \"80b77f5f-b572-4777-9912-800cc7886ddc\") " Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.003229 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80b77f5f-b572-4777-9912-800cc7886ddc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "80b77f5f-b572-4777-9912-800cc7886ddc" (UID: "80b77f5f-b572-4777-9912-800cc7886ddc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.003380 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80b77f5f-b572-4777-9912-800cc7886ddc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "80b77f5f-b572-4777-9912-800cc7886ddc" (UID: "80b77f5f-b572-4777-9912-800cc7886ddc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.003522 4959 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80b77f5f-b572-4777-9912-800cc7886ddc-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.007902 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-scripts" (OuterVolumeSpecName: "scripts") pod "80b77f5f-b572-4777-9912-800cc7886ddc" (UID: "80b77f5f-b572-4777-9912-800cc7886ddc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.007905 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80b77f5f-b572-4777-9912-800cc7886ddc-kube-api-access-bwqsr" (OuterVolumeSpecName: "kube-api-access-bwqsr") pod "80b77f5f-b572-4777-9912-800cc7886ddc" (UID: "80b77f5f-b572-4777-9912-800cc7886ddc"). InnerVolumeSpecName "kube-api-access-bwqsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.029707 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "80b77f5f-b572-4777-9912-800cc7886ddc" (UID: "80b77f5f-b572-4777-9912-800cc7886ddc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.052346 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "80b77f5f-b572-4777-9912-800cc7886ddc" (UID: "80b77f5f-b572-4777-9912-800cc7886ddc"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.085860 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "80b77f5f-b572-4777-9912-800cc7886ddc" (UID: "80b77f5f-b572-4777-9912-800cc7886ddc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.106583 4959 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/80b77f5f-b572-4777-9912-800cc7886ddc-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.106613 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.106645 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.106655 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwqsr\" (UniqueName: \"kubernetes.io/projected/80b77f5f-b572-4777-9912-800cc7886ddc-kube-api-access-bwqsr\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.106664 4959 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.106672 4959 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.136768 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-config-data" (OuterVolumeSpecName: "config-data") pod "80b77f5f-b572-4777-9912-800cc7886ddc" (UID: "80b77f5f-b572-4777-9912-800cc7886ddc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.208278 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80b77f5f-b572-4777-9912-800cc7886ddc-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.803622 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"80b77f5f-b572-4777-9912-800cc7886ddc","Type":"ContainerDied","Data":"b99fe4719591640961f1eaca17de15e2e1481896e55e5b34efd6a6c76158b201"} Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.803673 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.803681 4959 scope.go:117] "RemoveContainer" containerID="a9eb1598f0d4f84a7bfc717906de92ea9e5a67b5f8a526d14bc6cb44f27d0100" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.845258 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.848853 4959 scope.go:117] "RemoveContainer" containerID="394f73fb3149de229dcb4af397e77e4a465d315ea05e727bfa839d6025fcdb05" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.852948 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.863509 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:06:03 crc kubenswrapper[4959]: E1007 14:06:03.864040 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b77f5f-b572-4777-9912-800cc7886ddc" containerName="proxy-httpd" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.864117 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b77f5f-b572-4777-9912-800cc7886ddc" containerName="proxy-httpd" Oct 07 14:06:03 crc kubenswrapper[4959]: E1007 14:06:03.864190 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b77f5f-b572-4777-9912-800cc7886ddc" containerName="ceilometer-notification-agent" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.864276 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b77f5f-b572-4777-9912-800cc7886ddc" containerName="ceilometer-notification-agent" Oct 07 14:06:03 crc kubenswrapper[4959]: E1007 14:06:03.864329 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b77f5f-b572-4777-9912-800cc7886ddc" containerName="ceilometer-central-agent" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.864384 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b77f5f-b572-4777-9912-800cc7886ddc" containerName="ceilometer-central-agent" Oct 07 14:06:03 crc kubenswrapper[4959]: E1007 14:06:03.864445 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b77f5f-b572-4777-9912-800cc7886ddc" containerName="sg-core" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.864493 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b77f5f-b572-4777-9912-800cc7886ddc" containerName="sg-core" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.864705 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="80b77f5f-b572-4777-9912-800cc7886ddc" containerName="proxy-httpd" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.864770 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="80b77f5f-b572-4777-9912-800cc7886ddc" containerName="ceilometer-notification-agent" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.864830 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="80b77f5f-b572-4777-9912-800cc7886ddc" containerName="sg-core" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.864884 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="80b77f5f-b572-4777-9912-800cc7886ddc" containerName="ceilometer-central-agent" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.866526 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.869582 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.869744 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.870352 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.874434 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.880724 4959 scope.go:117] "RemoveContainer" containerID="d083fcb14061bd5b6d3657744d575379b4a02f239d3e5703ba088b08b6136427" Oct 07 14:06:03 crc kubenswrapper[4959]: I1007 14:06:03.912445 4959 scope.go:117] "RemoveContainer" containerID="c1cacf38716e17dbbf7bd68462d448bdecbf3780bc758d010f17d82736b5c3ae" Oct 07 14:06:04 crc kubenswrapper[4959]: I1007 14:06:04.021236 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " pod="openstack/ceilometer-0" Oct 07 14:06:04 crc kubenswrapper[4959]: I1007 14:06:04.021289 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd6ea2ae-da94-4567-b894-187901295d11-log-httpd\") pod \"ceilometer-0\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " pod="openstack/ceilometer-0" Oct 07 14:06:04 crc kubenswrapper[4959]: I1007 14:06:04.021351 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-scripts\") pod \"ceilometer-0\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " pod="openstack/ceilometer-0" Oct 07 14:06:04 crc kubenswrapper[4959]: I1007 14:06:04.021389 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " pod="openstack/ceilometer-0" Oct 07 14:06:04 crc kubenswrapper[4959]: I1007 14:06:04.021421 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-config-data\") pod \"ceilometer-0\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " pod="openstack/ceilometer-0" Oct 07 14:06:04 crc kubenswrapper[4959]: I1007 14:06:04.021666 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thtvv\" (UniqueName: \"kubernetes.io/projected/bd6ea2ae-da94-4567-b894-187901295d11-kube-api-access-thtvv\") pod \"ceilometer-0\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " pod="openstack/ceilometer-0" Oct 07 14:06:04 crc kubenswrapper[4959]: I1007 14:06:04.021734 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " pod="openstack/ceilometer-0" Oct 07 14:06:04 crc kubenswrapper[4959]: I1007 14:06:04.021851 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd6ea2ae-da94-4567-b894-187901295d11-run-httpd\") pod \"ceilometer-0\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " pod="openstack/ceilometer-0" Oct 07 14:06:04 crc kubenswrapper[4959]: I1007 14:06:04.123730 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-config-data\") pod \"ceilometer-0\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " pod="openstack/ceilometer-0" Oct 07 14:06:04 crc kubenswrapper[4959]: I1007 14:06:04.123792 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thtvv\" (UniqueName: \"kubernetes.io/projected/bd6ea2ae-da94-4567-b894-187901295d11-kube-api-access-thtvv\") pod \"ceilometer-0\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " pod="openstack/ceilometer-0" Oct 07 14:06:04 crc kubenswrapper[4959]: I1007 14:06:04.123819 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " pod="openstack/ceilometer-0" Oct 07 14:06:04 crc kubenswrapper[4959]: I1007 14:06:04.123841 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd6ea2ae-da94-4567-b894-187901295d11-run-httpd\") pod \"ceilometer-0\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " pod="openstack/ceilometer-0" Oct 07 14:06:04 crc kubenswrapper[4959]: I1007 14:06:04.123902 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " pod="openstack/ceilometer-0" Oct 07 14:06:04 crc kubenswrapper[4959]: I1007 14:06:04.123920 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd6ea2ae-da94-4567-b894-187901295d11-log-httpd\") pod \"ceilometer-0\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " pod="openstack/ceilometer-0" Oct 07 14:06:04 crc kubenswrapper[4959]: I1007 14:06:04.123960 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-scripts\") pod \"ceilometer-0\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " pod="openstack/ceilometer-0" Oct 07 14:06:04 crc kubenswrapper[4959]: I1007 14:06:04.123986 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " pod="openstack/ceilometer-0" Oct 07 14:06:04 crc kubenswrapper[4959]: I1007 14:06:04.124966 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd6ea2ae-da94-4567-b894-187901295d11-run-httpd\") pod \"ceilometer-0\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " pod="openstack/ceilometer-0" Oct 07 14:06:04 crc kubenswrapper[4959]: I1007 14:06:04.125360 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd6ea2ae-da94-4567-b894-187901295d11-log-httpd\") pod \"ceilometer-0\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " pod="openstack/ceilometer-0" Oct 07 14:06:04 crc kubenswrapper[4959]: I1007 14:06:04.132810 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " pod="openstack/ceilometer-0" Oct 07 14:06:04 crc kubenswrapper[4959]: I1007 14:06:04.132853 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-scripts\") pod \"ceilometer-0\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " pod="openstack/ceilometer-0" Oct 07 14:06:04 crc kubenswrapper[4959]: I1007 14:06:04.133580 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " pod="openstack/ceilometer-0" Oct 07 14:06:04 crc kubenswrapper[4959]: I1007 14:06:04.134453 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " pod="openstack/ceilometer-0" Oct 07 14:06:04 crc kubenswrapper[4959]: I1007 14:06:04.136274 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-config-data\") pod \"ceilometer-0\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " pod="openstack/ceilometer-0" Oct 07 14:06:04 crc kubenswrapper[4959]: I1007 14:06:04.153693 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thtvv\" (UniqueName: \"kubernetes.io/projected/bd6ea2ae-da94-4567-b894-187901295d11-kube-api-access-thtvv\") pod \"ceilometer-0\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " pod="openstack/ceilometer-0" Oct 07 14:06:04 crc kubenswrapper[4959]: I1007 14:06:04.184305 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:06:04 crc kubenswrapper[4959]: W1007 14:06:04.643973 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd6ea2ae_da94_4567_b894_187901295d11.slice/crio-fbb0afb79f13cf9610b5fc3c35ae9de4ad08e1dc38695cd46f121ae59a40dc17 WatchSource:0}: Error finding container fbb0afb79f13cf9610b5fc3c35ae9de4ad08e1dc38695cd46f121ae59a40dc17: Status 404 returned error can't find the container with id fbb0afb79f13cf9610b5fc3c35ae9de4ad08e1dc38695cd46f121ae59a40dc17 Oct 07 14:06:04 crc kubenswrapper[4959]: I1007 14:06:04.645975 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:06:04 crc kubenswrapper[4959]: I1007 14:06:04.647497 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 14:06:04 crc kubenswrapper[4959]: I1007 14:06:04.662434 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80b77f5f-b572-4777-9912-800cc7886ddc" path="/var/lib/kubelet/pods/80b77f5f-b572-4777-9912-800cc7886ddc/volumes" Oct 07 14:06:04 crc kubenswrapper[4959]: I1007 14:06:04.813659 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd6ea2ae-da94-4567-b894-187901295d11","Type":"ContainerStarted","Data":"fbb0afb79f13cf9610b5fc3c35ae9de4ad08e1dc38695cd46f121ae59a40dc17"} Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.314956 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.351949 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.391614 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.448729 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8ce7f2f-cbb7-4daa-9955-d71c900e32d5-logs\") pod \"a8ce7f2f-cbb7-4daa-9955-d71c900e32d5\" (UID: \"a8ce7f2f-cbb7-4daa-9955-d71c900e32d5\") " Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.448833 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdvpz\" (UniqueName: \"kubernetes.io/projected/a8ce7f2f-cbb7-4daa-9955-d71c900e32d5-kube-api-access-zdvpz\") pod \"a8ce7f2f-cbb7-4daa-9955-d71c900e32d5\" (UID: \"a8ce7f2f-cbb7-4daa-9955-d71c900e32d5\") " Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.448881 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8ce7f2f-cbb7-4daa-9955-d71c900e32d5-combined-ca-bundle\") pod \"a8ce7f2f-cbb7-4daa-9955-d71c900e32d5\" (UID: \"a8ce7f2f-cbb7-4daa-9955-d71c900e32d5\") " Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.448986 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8ce7f2f-cbb7-4daa-9955-d71c900e32d5-config-data\") pod \"a8ce7f2f-cbb7-4daa-9955-d71c900e32d5\" (UID: \"a8ce7f2f-cbb7-4daa-9955-d71c900e32d5\") " Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.450382 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8ce7f2f-cbb7-4daa-9955-d71c900e32d5-logs" (OuterVolumeSpecName: "logs") pod "a8ce7f2f-cbb7-4daa-9955-d71c900e32d5" (UID: "a8ce7f2f-cbb7-4daa-9955-d71c900e32d5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.454617 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8ce7f2f-cbb7-4daa-9955-d71c900e32d5-kube-api-access-zdvpz" (OuterVolumeSpecName: "kube-api-access-zdvpz") pod "a8ce7f2f-cbb7-4daa-9955-d71c900e32d5" (UID: "a8ce7f2f-cbb7-4daa-9955-d71c900e32d5"). InnerVolumeSpecName "kube-api-access-zdvpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.492645 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8ce7f2f-cbb7-4daa-9955-d71c900e32d5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a8ce7f2f-cbb7-4daa-9955-d71c900e32d5" (UID: "a8ce7f2f-cbb7-4daa-9955-d71c900e32d5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.493685 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8ce7f2f-cbb7-4daa-9955-d71c900e32d5-config-data" (OuterVolumeSpecName: "config-data") pod "a8ce7f2f-cbb7-4daa-9955-d71c900e32d5" (UID: "a8ce7f2f-cbb7-4daa-9955-d71c900e32d5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.554792 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8ce7f2f-cbb7-4daa-9955-d71c900e32d5-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.554832 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8ce7f2f-cbb7-4daa-9955-d71c900e32d5-logs\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.554845 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdvpz\" (UniqueName: \"kubernetes.io/projected/a8ce7f2f-cbb7-4daa-9955-d71c900e32d5-kube-api-access-zdvpz\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.554860 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8ce7f2f-cbb7-4daa-9955-d71c900e32d5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.822351 4959 generic.go:334] "Generic (PLEG): container finished" podID="a8ce7f2f-cbb7-4daa-9955-d71c900e32d5" containerID="faa17a2a452375a503f4d9e6af1a500f6c57913d483377e6df70906306f2892d" exitCode=0 Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.822807 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a8ce7f2f-cbb7-4daa-9955-d71c900e32d5","Type":"ContainerDied","Data":"faa17a2a452375a503f4d9e6af1a500f6c57913d483377e6df70906306f2892d"} Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.822834 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a8ce7f2f-cbb7-4daa-9955-d71c900e32d5","Type":"ContainerDied","Data":"44acfb12438fe547c9f197f7df293407a059f6d424e32f0609b7673c5910b433"} Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.822851 4959 scope.go:117] "RemoveContainer" containerID="faa17a2a452375a503f4d9e6af1a500f6c57913d483377e6df70906306f2892d" Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.822991 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.827909 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd6ea2ae-da94-4567-b894-187901295d11","Type":"ContainerStarted","Data":"4356eb32c5bc35fe66019602e00451bd08bff3a537274c3dc6a8c211974bc33d"} Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.849774 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.857204 4959 scope.go:117] "RemoveContainer" containerID="0ff620573f010a5a7e6cd46a4ed873136ab51b6ecaaf0a166bece48e3f98c4a5" Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.863422 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.871941 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.891524 4959 scope.go:117] "RemoveContainer" containerID="faa17a2a452375a503f4d9e6af1a500f6c57913d483377e6df70906306f2892d" Oct 07 14:06:05 crc kubenswrapper[4959]: E1007 14:06:05.892007 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"faa17a2a452375a503f4d9e6af1a500f6c57913d483377e6df70906306f2892d\": container with ID starting with faa17a2a452375a503f4d9e6af1a500f6c57913d483377e6df70906306f2892d not found: ID does not exist" containerID="faa17a2a452375a503f4d9e6af1a500f6c57913d483377e6df70906306f2892d" Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.892036 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"faa17a2a452375a503f4d9e6af1a500f6c57913d483377e6df70906306f2892d"} err="failed to get container status \"faa17a2a452375a503f4d9e6af1a500f6c57913d483377e6df70906306f2892d\": rpc error: code = NotFound desc = could not find container \"faa17a2a452375a503f4d9e6af1a500f6c57913d483377e6df70906306f2892d\": container with ID starting with faa17a2a452375a503f4d9e6af1a500f6c57913d483377e6df70906306f2892d not found: ID does not exist" Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.892057 4959 scope.go:117] "RemoveContainer" containerID="0ff620573f010a5a7e6cd46a4ed873136ab51b6ecaaf0a166bece48e3f98c4a5" Oct 07 14:06:05 crc kubenswrapper[4959]: E1007 14:06:05.892375 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ff620573f010a5a7e6cd46a4ed873136ab51b6ecaaf0a166bece48e3f98c4a5\": container with ID starting with 0ff620573f010a5a7e6cd46a4ed873136ab51b6ecaaf0a166bece48e3f98c4a5 not found: ID does not exist" containerID="0ff620573f010a5a7e6cd46a4ed873136ab51b6ecaaf0a166bece48e3f98c4a5" Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.892398 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ff620573f010a5a7e6cd46a4ed873136ab51b6ecaaf0a166bece48e3f98c4a5"} err="failed to get container status \"0ff620573f010a5a7e6cd46a4ed873136ab51b6ecaaf0a166bece48e3f98c4a5\": rpc error: code = NotFound desc = could not find container \"0ff620573f010a5a7e6cd46a4ed873136ab51b6ecaaf0a166bece48e3f98c4a5\": container with ID starting with 0ff620573f010a5a7e6cd46a4ed873136ab51b6ecaaf0a166bece48e3f98c4a5 not found: ID does not exist" Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.896940 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 07 14:06:05 crc kubenswrapper[4959]: E1007 14:06:05.897293 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8ce7f2f-cbb7-4daa-9955-d71c900e32d5" containerName="nova-api-log" Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.897312 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8ce7f2f-cbb7-4daa-9955-d71c900e32d5" containerName="nova-api-log" Oct 07 14:06:05 crc kubenswrapper[4959]: E1007 14:06:05.897341 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8ce7f2f-cbb7-4daa-9955-d71c900e32d5" containerName="nova-api-api" Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.897348 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8ce7f2f-cbb7-4daa-9955-d71c900e32d5" containerName="nova-api-api" Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.897505 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8ce7f2f-cbb7-4daa-9955-d71c900e32d5" containerName="nova-api-log" Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.897525 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8ce7f2f-cbb7-4daa-9955-d71c900e32d5" containerName="nova-api-api" Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.898402 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.901806 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.901963 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.902114 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 07 14:06:05 crc kubenswrapper[4959]: I1007 14:06:05.904517 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.013635 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-dkgbw"] Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.014733 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dkgbw" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.018991 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.019219 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.028635 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-dkgbw"] Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.063475 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wpwc\" (UniqueName: \"kubernetes.io/projected/f1daafef-ea7f-42be-adf2-3fe4a6003995-kube-api-access-6wpwc\") pod \"nova-api-0\" (UID: \"f1daafef-ea7f-42be-adf2-3fe4a6003995\") " pod="openstack/nova-api-0" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.063580 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1daafef-ea7f-42be-adf2-3fe4a6003995-public-tls-certs\") pod \"nova-api-0\" (UID: \"f1daafef-ea7f-42be-adf2-3fe4a6003995\") " pod="openstack/nova-api-0" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.063637 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1daafef-ea7f-42be-adf2-3fe4a6003995-config-data\") pod \"nova-api-0\" (UID: \"f1daafef-ea7f-42be-adf2-3fe4a6003995\") " pod="openstack/nova-api-0" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.063653 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1daafef-ea7f-42be-adf2-3fe4a6003995-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f1daafef-ea7f-42be-adf2-3fe4a6003995\") " pod="openstack/nova-api-0" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.063682 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1daafef-ea7f-42be-adf2-3fe4a6003995-logs\") pod \"nova-api-0\" (UID: \"f1daafef-ea7f-42be-adf2-3fe4a6003995\") " pod="openstack/nova-api-0" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.063698 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1daafef-ea7f-42be-adf2-3fe4a6003995-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f1daafef-ea7f-42be-adf2-3fe4a6003995\") " pod="openstack/nova-api-0" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.164788 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1daafef-ea7f-42be-adf2-3fe4a6003995-public-tls-certs\") pod \"nova-api-0\" (UID: \"f1daafef-ea7f-42be-adf2-3fe4a6003995\") " pod="openstack/nova-api-0" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.164921 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1daafef-ea7f-42be-adf2-3fe4a6003995-config-data\") pod \"nova-api-0\" (UID: \"f1daafef-ea7f-42be-adf2-3fe4a6003995\") " pod="openstack/nova-api-0" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.164949 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1daafef-ea7f-42be-adf2-3fe4a6003995-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f1daafef-ea7f-42be-adf2-3fe4a6003995\") " pod="openstack/nova-api-0" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.165034 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1daafef-ea7f-42be-adf2-3fe4a6003995-logs\") pod \"nova-api-0\" (UID: \"f1daafef-ea7f-42be-adf2-3fe4a6003995\") " pod="openstack/nova-api-0" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.165053 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1daafef-ea7f-42be-adf2-3fe4a6003995-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f1daafef-ea7f-42be-adf2-3fe4a6003995\") " pod="openstack/nova-api-0" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.165132 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kb9l\" (UniqueName: \"kubernetes.io/projected/adff668e-a674-4ad1-ae8f-a28b2f6bc9e3-kube-api-access-7kb9l\") pod \"nova-cell1-cell-mapping-dkgbw\" (UID: \"adff668e-a674-4ad1-ae8f-a28b2f6bc9e3\") " pod="openstack/nova-cell1-cell-mapping-dkgbw" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.165179 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wpwc\" (UniqueName: \"kubernetes.io/projected/f1daafef-ea7f-42be-adf2-3fe4a6003995-kube-api-access-6wpwc\") pod \"nova-api-0\" (UID: \"f1daafef-ea7f-42be-adf2-3fe4a6003995\") " pod="openstack/nova-api-0" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.165230 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adff668e-a674-4ad1-ae8f-a28b2f6bc9e3-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dkgbw\" (UID: \"adff668e-a674-4ad1-ae8f-a28b2f6bc9e3\") " pod="openstack/nova-cell1-cell-mapping-dkgbw" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.165254 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adff668e-a674-4ad1-ae8f-a28b2f6bc9e3-scripts\") pod \"nova-cell1-cell-mapping-dkgbw\" (UID: \"adff668e-a674-4ad1-ae8f-a28b2f6bc9e3\") " pod="openstack/nova-cell1-cell-mapping-dkgbw" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.165352 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adff668e-a674-4ad1-ae8f-a28b2f6bc9e3-config-data\") pod \"nova-cell1-cell-mapping-dkgbw\" (UID: \"adff668e-a674-4ad1-ae8f-a28b2f6bc9e3\") " pod="openstack/nova-cell1-cell-mapping-dkgbw" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.165469 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1daafef-ea7f-42be-adf2-3fe4a6003995-logs\") pod \"nova-api-0\" (UID: \"f1daafef-ea7f-42be-adf2-3fe4a6003995\") " pod="openstack/nova-api-0" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.170060 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1daafef-ea7f-42be-adf2-3fe4a6003995-config-data\") pod \"nova-api-0\" (UID: \"f1daafef-ea7f-42be-adf2-3fe4a6003995\") " pod="openstack/nova-api-0" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.171784 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1daafef-ea7f-42be-adf2-3fe4a6003995-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f1daafef-ea7f-42be-adf2-3fe4a6003995\") " pod="openstack/nova-api-0" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.174613 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1daafef-ea7f-42be-adf2-3fe4a6003995-public-tls-certs\") pod \"nova-api-0\" (UID: \"f1daafef-ea7f-42be-adf2-3fe4a6003995\") " pod="openstack/nova-api-0" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.175071 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1daafef-ea7f-42be-adf2-3fe4a6003995-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f1daafef-ea7f-42be-adf2-3fe4a6003995\") " pod="openstack/nova-api-0" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.187142 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wpwc\" (UniqueName: \"kubernetes.io/projected/f1daafef-ea7f-42be-adf2-3fe4a6003995-kube-api-access-6wpwc\") pod \"nova-api-0\" (UID: \"f1daafef-ea7f-42be-adf2-3fe4a6003995\") " pod="openstack/nova-api-0" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.217906 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.267154 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kb9l\" (UniqueName: \"kubernetes.io/projected/adff668e-a674-4ad1-ae8f-a28b2f6bc9e3-kube-api-access-7kb9l\") pod \"nova-cell1-cell-mapping-dkgbw\" (UID: \"adff668e-a674-4ad1-ae8f-a28b2f6bc9e3\") " pod="openstack/nova-cell1-cell-mapping-dkgbw" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.267218 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adff668e-a674-4ad1-ae8f-a28b2f6bc9e3-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dkgbw\" (UID: \"adff668e-a674-4ad1-ae8f-a28b2f6bc9e3\") " pod="openstack/nova-cell1-cell-mapping-dkgbw" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.267242 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adff668e-a674-4ad1-ae8f-a28b2f6bc9e3-scripts\") pod \"nova-cell1-cell-mapping-dkgbw\" (UID: \"adff668e-a674-4ad1-ae8f-a28b2f6bc9e3\") " pod="openstack/nova-cell1-cell-mapping-dkgbw" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.267278 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adff668e-a674-4ad1-ae8f-a28b2f6bc9e3-config-data\") pod \"nova-cell1-cell-mapping-dkgbw\" (UID: \"adff668e-a674-4ad1-ae8f-a28b2f6bc9e3\") " pod="openstack/nova-cell1-cell-mapping-dkgbw" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.271919 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adff668e-a674-4ad1-ae8f-a28b2f6bc9e3-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dkgbw\" (UID: \"adff668e-a674-4ad1-ae8f-a28b2f6bc9e3\") " pod="openstack/nova-cell1-cell-mapping-dkgbw" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.272396 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adff668e-a674-4ad1-ae8f-a28b2f6bc9e3-scripts\") pod \"nova-cell1-cell-mapping-dkgbw\" (UID: \"adff668e-a674-4ad1-ae8f-a28b2f6bc9e3\") " pod="openstack/nova-cell1-cell-mapping-dkgbw" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.272799 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adff668e-a674-4ad1-ae8f-a28b2f6bc9e3-config-data\") pod \"nova-cell1-cell-mapping-dkgbw\" (UID: \"adff668e-a674-4ad1-ae8f-a28b2f6bc9e3\") " pod="openstack/nova-cell1-cell-mapping-dkgbw" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.288606 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kb9l\" (UniqueName: \"kubernetes.io/projected/adff668e-a674-4ad1-ae8f-a28b2f6bc9e3-kube-api-access-7kb9l\") pod \"nova-cell1-cell-mapping-dkgbw\" (UID: \"adff668e-a674-4ad1-ae8f-a28b2f6bc9e3\") " pod="openstack/nova-cell1-cell-mapping-dkgbw" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.335050 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dkgbw" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.663849 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8ce7f2f-cbb7-4daa-9955-d71c900e32d5" path="/var/lib/kubelet/pods/a8ce7f2f-cbb7-4daa-9955-d71c900e32d5/volumes" Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.716417 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 14:06:06 crc kubenswrapper[4959]: W1007 14:06:06.716695 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1daafef_ea7f_42be_adf2_3fe4a6003995.slice/crio-7767520d729462f27ca80c12c8ddfad0611ac1c820b6a67237943d7725f27ec3 WatchSource:0}: Error finding container 7767520d729462f27ca80c12c8ddfad0611ac1c820b6a67237943d7725f27ec3: Status 404 returned error can't find the container with id 7767520d729462f27ca80c12c8ddfad0611ac1c820b6a67237943d7725f27ec3 Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.799154 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-dkgbw"] Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.841139 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dkgbw" event={"ID":"adff668e-a674-4ad1-ae8f-a28b2f6bc9e3","Type":"ContainerStarted","Data":"88777ce3d54d000a2f0a589c2db4a13ec79c0772a3c23737f2802beca4b1a112"} Oct 07 14:06:06 crc kubenswrapper[4959]: I1007 14:06:06.844979 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f1daafef-ea7f-42be-adf2-3fe4a6003995","Type":"ContainerStarted","Data":"7767520d729462f27ca80c12c8ddfad0611ac1c820b6a67237943d7725f27ec3"} Oct 07 14:06:07 crc kubenswrapper[4959]: I1007 14:06:07.855339 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dkgbw" event={"ID":"adff668e-a674-4ad1-ae8f-a28b2f6bc9e3","Type":"ContainerStarted","Data":"034f3f30a554074bc27c3bf920535bc8bd69bcef6bf3b7bfc5419fa2bc4ed263"} Oct 07 14:06:07 crc kubenswrapper[4959]: I1007 14:06:07.857603 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f1daafef-ea7f-42be-adf2-3fe4a6003995","Type":"ContainerStarted","Data":"e75946aa0c81914320f17158bf624ef5cdb2122addafae9457ba63d3cd2c8c88"} Oct 07 14:06:07 crc kubenswrapper[4959]: I1007 14:06:07.857651 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f1daafef-ea7f-42be-adf2-3fe4a6003995","Type":"ContainerStarted","Data":"cdef07dc561bd028518e79c39104ade93f6359395da1f024ed411caffa758ecc"} Oct 07 14:06:07 crc kubenswrapper[4959]: I1007 14:06:07.879240 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-dkgbw" podStartSLOduration=2.879216179 podStartE2EDuration="2.879216179s" podCreationTimestamp="2025-10-07 14:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:06:07.877605939 +0000 UTC m=+1249.961010264" watchObservedRunningTime="2025-10-07 14:06:07.879216179 +0000 UTC m=+1249.962620504" Oct 07 14:06:08 crc kubenswrapper[4959]: I1007 14:06:08.686266 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.686245283 podStartE2EDuration="3.686245283s" podCreationTimestamp="2025-10-07 14:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:06:07.899461437 +0000 UTC m=+1249.982865762" watchObservedRunningTime="2025-10-07 14:06:08.686245283 +0000 UTC m=+1250.769649608" Oct 07 14:06:08 crc kubenswrapper[4959]: I1007 14:06:08.868548 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd6ea2ae-da94-4567-b894-187901295d11","Type":"ContainerStarted","Data":"facb0655c5f056bc6a8956fc3e14daa62f6d0820180538b3f17c1e1e4a630a76"} Oct 07 14:06:08 crc kubenswrapper[4959]: I1007 14:06:08.868591 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd6ea2ae-da94-4567-b894-187901295d11","Type":"ContainerStarted","Data":"ee10c2c91d1929fe62bea6ab20cac73239dd77669957a696ebe922a81a9c40a4"} Oct 07 14:06:09 crc kubenswrapper[4959]: I1007 14:06:09.234240 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-665946c669-2wpls" Oct 07 14:06:09 crc kubenswrapper[4959]: I1007 14:06:09.321355 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75fb48c489-xpjqw"] Oct 07 14:06:09 crc kubenswrapper[4959]: I1007 14:06:09.321879 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75fb48c489-xpjqw" podUID="7fc43cc8-61cc-4102-aacc-356966a9b9a5" containerName="dnsmasq-dns" containerID="cri-o://037d18a203529a0c452135fb14b887cb6170bcdf0231d57206c64653e800d31e" gracePeriod=10 Oct 07 14:06:09 crc kubenswrapper[4959]: I1007 14:06:09.889137 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75fb48c489-xpjqw" Oct 07 14:06:09 crc kubenswrapper[4959]: I1007 14:06:09.891916 4959 generic.go:334] "Generic (PLEG): container finished" podID="7fc43cc8-61cc-4102-aacc-356966a9b9a5" containerID="037d18a203529a0c452135fb14b887cb6170bcdf0231d57206c64653e800d31e" exitCode=0 Oct 07 14:06:09 crc kubenswrapper[4959]: I1007 14:06:09.891999 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75fb48c489-xpjqw" event={"ID":"7fc43cc8-61cc-4102-aacc-356966a9b9a5","Type":"ContainerDied","Data":"037d18a203529a0c452135fb14b887cb6170bcdf0231d57206c64653e800d31e"} Oct 07 14:06:09 crc kubenswrapper[4959]: I1007 14:06:09.892029 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75fb48c489-xpjqw" event={"ID":"7fc43cc8-61cc-4102-aacc-356966a9b9a5","Type":"ContainerDied","Data":"fd25e7a57f9d5fff59528486ff53d97d7645e6265dd4fe52648bbf1cc594f1a4"} Oct 07 14:06:09 crc kubenswrapper[4959]: I1007 14:06:09.892076 4959 scope.go:117] "RemoveContainer" containerID="037d18a203529a0c452135fb14b887cb6170bcdf0231d57206c64653e800d31e" Oct 07 14:06:09 crc kubenswrapper[4959]: I1007 14:06:09.931430 4959 scope.go:117] "RemoveContainer" containerID="9105e8d5198acec26c4d4032d4cc9fe24be4b3e3b51be2f34f12c879c005373c" Oct 07 14:06:09 crc kubenswrapper[4959]: I1007 14:06:09.963809 4959 scope.go:117] "RemoveContainer" containerID="037d18a203529a0c452135fb14b887cb6170bcdf0231d57206c64653e800d31e" Oct 07 14:06:09 crc kubenswrapper[4959]: E1007 14:06:09.966598 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"037d18a203529a0c452135fb14b887cb6170bcdf0231d57206c64653e800d31e\": container with ID starting with 037d18a203529a0c452135fb14b887cb6170bcdf0231d57206c64653e800d31e not found: ID does not exist" containerID="037d18a203529a0c452135fb14b887cb6170bcdf0231d57206c64653e800d31e" Oct 07 14:06:09 crc kubenswrapper[4959]: I1007 14:06:09.966639 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"037d18a203529a0c452135fb14b887cb6170bcdf0231d57206c64653e800d31e"} err="failed to get container status \"037d18a203529a0c452135fb14b887cb6170bcdf0231d57206c64653e800d31e\": rpc error: code = NotFound desc = could not find container \"037d18a203529a0c452135fb14b887cb6170bcdf0231d57206c64653e800d31e\": container with ID starting with 037d18a203529a0c452135fb14b887cb6170bcdf0231d57206c64653e800d31e not found: ID does not exist" Oct 07 14:06:09 crc kubenswrapper[4959]: I1007 14:06:09.966666 4959 scope.go:117] "RemoveContainer" containerID="9105e8d5198acec26c4d4032d4cc9fe24be4b3e3b51be2f34f12c879c005373c" Oct 07 14:06:09 crc kubenswrapper[4959]: E1007 14:06:09.967007 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9105e8d5198acec26c4d4032d4cc9fe24be4b3e3b51be2f34f12c879c005373c\": container with ID starting with 9105e8d5198acec26c4d4032d4cc9fe24be4b3e3b51be2f34f12c879c005373c not found: ID does not exist" containerID="9105e8d5198acec26c4d4032d4cc9fe24be4b3e3b51be2f34f12c879c005373c" Oct 07 14:06:09 crc kubenswrapper[4959]: I1007 14:06:09.967040 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9105e8d5198acec26c4d4032d4cc9fe24be4b3e3b51be2f34f12c879c005373c"} err="failed to get container status \"9105e8d5198acec26c4d4032d4cc9fe24be4b3e3b51be2f34f12c879c005373c\": rpc error: code = NotFound desc = could not find container \"9105e8d5198acec26c4d4032d4cc9fe24be4b3e3b51be2f34f12c879c005373c\": container with ID starting with 9105e8d5198acec26c4d4032d4cc9fe24be4b3e3b51be2f34f12c879c005373c not found: ID does not exist" Oct 07 14:06:10 crc kubenswrapper[4959]: I1007 14:06:10.048274 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fc43cc8-61cc-4102-aacc-356966a9b9a5-config\") pod \"7fc43cc8-61cc-4102-aacc-356966a9b9a5\" (UID: \"7fc43cc8-61cc-4102-aacc-356966a9b9a5\") " Oct 07 14:06:10 crc kubenswrapper[4959]: I1007 14:06:10.048415 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7fc43cc8-61cc-4102-aacc-356966a9b9a5-ovsdbserver-sb\") pod \"7fc43cc8-61cc-4102-aacc-356966a9b9a5\" (UID: \"7fc43cc8-61cc-4102-aacc-356966a9b9a5\") " Oct 07 14:06:10 crc kubenswrapper[4959]: I1007 14:06:10.048512 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fc43cc8-61cc-4102-aacc-356966a9b9a5-ovsdbserver-nb\") pod \"7fc43cc8-61cc-4102-aacc-356966a9b9a5\" (UID: \"7fc43cc8-61cc-4102-aacc-356966a9b9a5\") " Oct 07 14:06:10 crc kubenswrapper[4959]: I1007 14:06:10.048550 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fc43cc8-61cc-4102-aacc-356966a9b9a5-dns-svc\") pod \"7fc43cc8-61cc-4102-aacc-356966a9b9a5\" (UID: \"7fc43cc8-61cc-4102-aacc-356966a9b9a5\") " Oct 07 14:06:10 crc kubenswrapper[4959]: I1007 14:06:10.048601 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khxkr\" (UniqueName: \"kubernetes.io/projected/7fc43cc8-61cc-4102-aacc-356966a9b9a5-kube-api-access-khxkr\") pod \"7fc43cc8-61cc-4102-aacc-356966a9b9a5\" (UID: \"7fc43cc8-61cc-4102-aacc-356966a9b9a5\") " Oct 07 14:06:10 crc kubenswrapper[4959]: I1007 14:06:10.053381 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fc43cc8-61cc-4102-aacc-356966a9b9a5-kube-api-access-khxkr" (OuterVolumeSpecName: "kube-api-access-khxkr") pod "7fc43cc8-61cc-4102-aacc-356966a9b9a5" (UID: "7fc43cc8-61cc-4102-aacc-356966a9b9a5"). InnerVolumeSpecName "kube-api-access-khxkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:06:10 crc kubenswrapper[4959]: I1007 14:06:10.101973 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fc43cc8-61cc-4102-aacc-356966a9b9a5-config" (OuterVolumeSpecName: "config") pod "7fc43cc8-61cc-4102-aacc-356966a9b9a5" (UID: "7fc43cc8-61cc-4102-aacc-356966a9b9a5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:06:10 crc kubenswrapper[4959]: I1007 14:06:10.111054 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fc43cc8-61cc-4102-aacc-356966a9b9a5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7fc43cc8-61cc-4102-aacc-356966a9b9a5" (UID: "7fc43cc8-61cc-4102-aacc-356966a9b9a5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:06:10 crc kubenswrapper[4959]: I1007 14:06:10.122005 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fc43cc8-61cc-4102-aacc-356966a9b9a5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7fc43cc8-61cc-4102-aacc-356966a9b9a5" (UID: "7fc43cc8-61cc-4102-aacc-356966a9b9a5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:06:10 crc kubenswrapper[4959]: I1007 14:06:10.128445 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fc43cc8-61cc-4102-aacc-356966a9b9a5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7fc43cc8-61cc-4102-aacc-356966a9b9a5" (UID: "7fc43cc8-61cc-4102-aacc-356966a9b9a5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:06:10 crc kubenswrapper[4959]: I1007 14:06:10.152942 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fc43cc8-61cc-4102-aacc-356966a9b9a5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:10 crc kubenswrapper[4959]: I1007 14:06:10.153069 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fc43cc8-61cc-4102-aacc-356966a9b9a5-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:10 crc kubenswrapper[4959]: I1007 14:06:10.153165 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khxkr\" (UniqueName: \"kubernetes.io/projected/7fc43cc8-61cc-4102-aacc-356966a9b9a5-kube-api-access-khxkr\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:10 crc kubenswrapper[4959]: I1007 14:06:10.153243 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fc43cc8-61cc-4102-aacc-356966a9b9a5-config\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:10 crc kubenswrapper[4959]: I1007 14:06:10.153327 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7fc43cc8-61cc-4102-aacc-356966a9b9a5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:10 crc kubenswrapper[4959]: I1007 14:06:10.905601 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd6ea2ae-da94-4567-b894-187901295d11","Type":"ContainerStarted","Data":"1484be398cc30548478f33867e65fcedcc0aa3cfb54643f6125d3cf9499fae12"} Oct 07 14:06:10 crc kubenswrapper[4959]: I1007 14:06:10.906141 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 07 14:06:10 crc kubenswrapper[4959]: I1007 14:06:10.907574 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75fb48c489-xpjqw" Oct 07 14:06:10 crc kubenswrapper[4959]: I1007 14:06:10.933287 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.465279385 podStartE2EDuration="7.933268721s" podCreationTimestamp="2025-10-07 14:06:03 +0000 UTC" firstStartedPulling="2025-10-07 14:06:04.647161574 +0000 UTC m=+1246.730565899" lastFinishedPulling="2025-10-07 14:06:10.11515091 +0000 UTC m=+1252.198555235" observedRunningTime="2025-10-07 14:06:10.928243495 +0000 UTC m=+1253.011647850" watchObservedRunningTime="2025-10-07 14:06:10.933268721 +0000 UTC m=+1253.016673046" Oct 07 14:06:10 crc kubenswrapper[4959]: I1007 14:06:10.960839 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75fb48c489-xpjqw"] Oct 07 14:06:10 crc kubenswrapper[4959]: I1007 14:06:10.966840 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75fb48c489-xpjqw"] Oct 07 14:06:12 crc kubenswrapper[4959]: I1007 14:06:12.667085 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fc43cc8-61cc-4102-aacc-356966a9b9a5" path="/var/lib/kubelet/pods/7fc43cc8-61cc-4102-aacc-356966a9b9a5/volumes" Oct 07 14:06:12 crc kubenswrapper[4959]: I1007 14:06:12.924119 4959 generic.go:334] "Generic (PLEG): container finished" podID="adff668e-a674-4ad1-ae8f-a28b2f6bc9e3" containerID="034f3f30a554074bc27c3bf920535bc8bd69bcef6bf3b7bfc5419fa2bc4ed263" exitCode=0 Oct 07 14:06:12 crc kubenswrapper[4959]: I1007 14:06:12.924159 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dkgbw" event={"ID":"adff668e-a674-4ad1-ae8f-a28b2f6bc9e3","Type":"ContainerDied","Data":"034f3f30a554074bc27c3bf920535bc8bd69bcef6bf3b7bfc5419fa2bc4ed263"} Oct 07 14:06:14 crc kubenswrapper[4959]: I1007 14:06:14.342163 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dkgbw" Oct 07 14:06:14 crc kubenswrapper[4959]: I1007 14:06:14.529775 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adff668e-a674-4ad1-ae8f-a28b2f6bc9e3-config-data\") pod \"adff668e-a674-4ad1-ae8f-a28b2f6bc9e3\" (UID: \"adff668e-a674-4ad1-ae8f-a28b2f6bc9e3\") " Oct 07 14:06:14 crc kubenswrapper[4959]: I1007 14:06:14.529965 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adff668e-a674-4ad1-ae8f-a28b2f6bc9e3-scripts\") pod \"adff668e-a674-4ad1-ae8f-a28b2f6bc9e3\" (UID: \"adff668e-a674-4ad1-ae8f-a28b2f6bc9e3\") " Oct 07 14:06:14 crc kubenswrapper[4959]: I1007 14:06:14.529992 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kb9l\" (UniqueName: \"kubernetes.io/projected/adff668e-a674-4ad1-ae8f-a28b2f6bc9e3-kube-api-access-7kb9l\") pod \"adff668e-a674-4ad1-ae8f-a28b2f6bc9e3\" (UID: \"adff668e-a674-4ad1-ae8f-a28b2f6bc9e3\") " Oct 07 14:06:14 crc kubenswrapper[4959]: I1007 14:06:14.530061 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adff668e-a674-4ad1-ae8f-a28b2f6bc9e3-combined-ca-bundle\") pod \"adff668e-a674-4ad1-ae8f-a28b2f6bc9e3\" (UID: \"adff668e-a674-4ad1-ae8f-a28b2f6bc9e3\") " Oct 07 14:06:14 crc kubenswrapper[4959]: I1007 14:06:14.538116 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adff668e-a674-4ad1-ae8f-a28b2f6bc9e3-kube-api-access-7kb9l" (OuterVolumeSpecName: "kube-api-access-7kb9l") pod "adff668e-a674-4ad1-ae8f-a28b2f6bc9e3" (UID: "adff668e-a674-4ad1-ae8f-a28b2f6bc9e3"). InnerVolumeSpecName "kube-api-access-7kb9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:06:14 crc kubenswrapper[4959]: I1007 14:06:14.555093 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adff668e-a674-4ad1-ae8f-a28b2f6bc9e3-scripts" (OuterVolumeSpecName: "scripts") pod "adff668e-a674-4ad1-ae8f-a28b2f6bc9e3" (UID: "adff668e-a674-4ad1-ae8f-a28b2f6bc9e3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:06:14 crc kubenswrapper[4959]: I1007 14:06:14.578382 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adff668e-a674-4ad1-ae8f-a28b2f6bc9e3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "adff668e-a674-4ad1-ae8f-a28b2f6bc9e3" (UID: "adff668e-a674-4ad1-ae8f-a28b2f6bc9e3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:06:14 crc kubenswrapper[4959]: I1007 14:06:14.583860 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adff668e-a674-4ad1-ae8f-a28b2f6bc9e3-config-data" (OuterVolumeSpecName: "config-data") pod "adff668e-a674-4ad1-ae8f-a28b2f6bc9e3" (UID: "adff668e-a674-4ad1-ae8f-a28b2f6bc9e3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:06:14 crc kubenswrapper[4959]: I1007 14:06:14.632283 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adff668e-a674-4ad1-ae8f-a28b2f6bc9e3-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:14 crc kubenswrapper[4959]: I1007 14:06:14.632326 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kb9l\" (UniqueName: \"kubernetes.io/projected/adff668e-a674-4ad1-ae8f-a28b2f6bc9e3-kube-api-access-7kb9l\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:14 crc kubenswrapper[4959]: I1007 14:06:14.632341 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adff668e-a674-4ad1-ae8f-a28b2f6bc9e3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:14 crc kubenswrapper[4959]: I1007 14:06:14.632375 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adff668e-a674-4ad1-ae8f-a28b2f6bc9e3-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:14 crc kubenswrapper[4959]: I1007 14:06:14.700334 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-75fb48c489-xpjqw" podUID="7fc43cc8-61cc-4102-aacc-356966a9b9a5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.181:5353: i/o timeout" Oct 07 14:06:14 crc kubenswrapper[4959]: I1007 14:06:14.954170 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dkgbw" event={"ID":"adff668e-a674-4ad1-ae8f-a28b2f6bc9e3","Type":"ContainerDied","Data":"88777ce3d54d000a2f0a589c2db4a13ec79c0772a3c23737f2802beca4b1a112"} Oct 07 14:06:14 crc kubenswrapper[4959]: I1007 14:06:14.954498 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88777ce3d54d000a2f0a589c2db4a13ec79c0772a3c23737f2802beca4b1a112" Oct 07 14:06:14 crc kubenswrapper[4959]: I1007 14:06:14.954552 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dkgbw" Oct 07 14:06:15 crc kubenswrapper[4959]: I1007 14:06:15.119084 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 07 14:06:15 crc kubenswrapper[4959]: I1007 14:06:15.119390 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f1daafef-ea7f-42be-adf2-3fe4a6003995" containerName="nova-api-log" containerID="cri-o://cdef07dc561bd028518e79c39104ade93f6359395da1f024ed411caffa758ecc" gracePeriod=30 Oct 07 14:06:15 crc kubenswrapper[4959]: I1007 14:06:15.119507 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f1daafef-ea7f-42be-adf2-3fe4a6003995" containerName="nova-api-api" containerID="cri-o://e75946aa0c81914320f17158bf624ef5cdb2122addafae9457ba63d3cd2c8c88" gracePeriod=30 Oct 07 14:06:15 crc kubenswrapper[4959]: I1007 14:06:15.141823 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 14:06:15 crc kubenswrapper[4959]: I1007 14:06:15.142233 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="14f0652c-3a52-460c-9ebc-86eeba72e77a" containerName="nova-scheduler-scheduler" containerID="cri-o://48cf7552cac7401d7fb488ed8ed9ab98b4408a52de7d47873248702d94af9ad5" gracePeriod=30 Oct 07 14:06:15 crc kubenswrapper[4959]: I1007 14:06:15.220236 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 14:06:15 crc kubenswrapper[4959]: I1007 14:06:15.220464 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0a9c0a78-b2ae-4edb-8b49-be96b1b0731f" containerName="nova-metadata-log" containerID="cri-o://c24cd83763ffe59dcb7a1d47b5f5811d39efff6aeb0689d5ea52a3e526d61ba4" gracePeriod=30 Oct 07 14:06:15 crc kubenswrapper[4959]: I1007 14:06:15.220562 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0a9c0a78-b2ae-4edb-8b49-be96b1b0731f" containerName="nova-metadata-metadata" containerID="cri-o://8cd1268f6a16cc19070e1500a6303258aa9ca5e7bfa867d9350c4ac21ee4d585" gracePeriod=30 Oct 07 14:06:15 crc kubenswrapper[4959]: I1007 14:06:15.822992 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 14:06:15 crc kubenswrapper[4959]: I1007 14:06:15.961302 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1daafef-ea7f-42be-adf2-3fe4a6003995-internal-tls-certs\") pod \"f1daafef-ea7f-42be-adf2-3fe4a6003995\" (UID: \"f1daafef-ea7f-42be-adf2-3fe4a6003995\") " Oct 07 14:06:15 crc kubenswrapper[4959]: I1007 14:06:15.961408 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1daafef-ea7f-42be-adf2-3fe4a6003995-logs\") pod \"f1daafef-ea7f-42be-adf2-3fe4a6003995\" (UID: \"f1daafef-ea7f-42be-adf2-3fe4a6003995\") " Oct 07 14:06:15 crc kubenswrapper[4959]: I1007 14:06:15.961430 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wpwc\" (UniqueName: \"kubernetes.io/projected/f1daafef-ea7f-42be-adf2-3fe4a6003995-kube-api-access-6wpwc\") pod \"f1daafef-ea7f-42be-adf2-3fe4a6003995\" (UID: \"f1daafef-ea7f-42be-adf2-3fe4a6003995\") " Oct 07 14:06:15 crc kubenswrapper[4959]: I1007 14:06:15.961462 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1daafef-ea7f-42be-adf2-3fe4a6003995-combined-ca-bundle\") pod \"f1daafef-ea7f-42be-adf2-3fe4a6003995\" (UID: \"f1daafef-ea7f-42be-adf2-3fe4a6003995\") " Oct 07 14:06:15 crc kubenswrapper[4959]: I1007 14:06:15.961498 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1daafef-ea7f-42be-adf2-3fe4a6003995-public-tls-certs\") pod \"f1daafef-ea7f-42be-adf2-3fe4a6003995\" (UID: \"f1daafef-ea7f-42be-adf2-3fe4a6003995\") " Oct 07 14:06:15 crc kubenswrapper[4959]: I1007 14:06:15.961516 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1daafef-ea7f-42be-adf2-3fe4a6003995-config-data\") pod \"f1daafef-ea7f-42be-adf2-3fe4a6003995\" (UID: \"f1daafef-ea7f-42be-adf2-3fe4a6003995\") " Oct 07 14:06:15 crc kubenswrapper[4959]: I1007 14:06:15.961995 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1daafef-ea7f-42be-adf2-3fe4a6003995-logs" (OuterVolumeSpecName: "logs") pod "f1daafef-ea7f-42be-adf2-3fe4a6003995" (UID: "f1daafef-ea7f-42be-adf2-3fe4a6003995"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:06:15 crc kubenswrapper[4959]: I1007 14:06:15.968573 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1daafef-ea7f-42be-adf2-3fe4a6003995-kube-api-access-6wpwc" (OuterVolumeSpecName: "kube-api-access-6wpwc") pod "f1daafef-ea7f-42be-adf2-3fe4a6003995" (UID: "f1daafef-ea7f-42be-adf2-3fe4a6003995"). InnerVolumeSpecName "kube-api-access-6wpwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:06:15 crc kubenswrapper[4959]: I1007 14:06:15.990730 4959 generic.go:334] "Generic (PLEG): container finished" podID="f1daafef-ea7f-42be-adf2-3fe4a6003995" containerID="e75946aa0c81914320f17158bf624ef5cdb2122addafae9457ba63d3cd2c8c88" exitCode=0 Oct 07 14:06:15 crc kubenswrapper[4959]: I1007 14:06:15.990762 4959 generic.go:334] "Generic (PLEG): container finished" podID="f1daafef-ea7f-42be-adf2-3fe4a6003995" containerID="cdef07dc561bd028518e79c39104ade93f6359395da1f024ed411caffa758ecc" exitCode=143 Oct 07 14:06:15 crc kubenswrapper[4959]: I1007 14:06:15.990810 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f1daafef-ea7f-42be-adf2-3fe4a6003995","Type":"ContainerDied","Data":"e75946aa0c81914320f17158bf624ef5cdb2122addafae9457ba63d3cd2c8c88"} Oct 07 14:06:15 crc kubenswrapper[4959]: I1007 14:06:15.990894 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 14:06:15 crc kubenswrapper[4959]: I1007 14:06:15.991184 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f1daafef-ea7f-42be-adf2-3fe4a6003995","Type":"ContainerDied","Data":"cdef07dc561bd028518e79c39104ade93f6359395da1f024ed411caffa758ecc"} Oct 07 14:06:15 crc kubenswrapper[4959]: I1007 14:06:15.991280 4959 scope.go:117] "RemoveContainer" containerID="e75946aa0c81914320f17158bf624ef5cdb2122addafae9457ba63d3cd2c8c88" Oct 07 14:06:15 crc kubenswrapper[4959]: I1007 14:06:15.993941 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1daafef-ea7f-42be-adf2-3fe4a6003995-config-data" (OuterVolumeSpecName: "config-data") pod "f1daafef-ea7f-42be-adf2-3fe4a6003995" (UID: "f1daafef-ea7f-42be-adf2-3fe4a6003995"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:06:15 crc kubenswrapper[4959]: I1007 14:06:15.997266 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f1daafef-ea7f-42be-adf2-3fe4a6003995","Type":"ContainerDied","Data":"7767520d729462f27ca80c12c8ddfad0611ac1c820b6a67237943d7725f27ec3"} Oct 07 14:06:15 crc kubenswrapper[4959]: I1007 14:06:15.997334 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f","Type":"ContainerDied","Data":"c24cd83763ffe59dcb7a1d47b5f5811d39efff6aeb0689d5ea52a3e526d61ba4"} Oct 07 14:06:15 crc kubenswrapper[4959]: I1007 14:06:15.997363 4959 generic.go:334] "Generic (PLEG): container finished" podID="0a9c0a78-b2ae-4edb-8b49-be96b1b0731f" containerID="c24cd83763ffe59dcb7a1d47b5f5811d39efff6aeb0689d5ea52a3e526d61ba4" exitCode=143 Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.000879 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1daafef-ea7f-42be-adf2-3fe4a6003995-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f1daafef-ea7f-42be-adf2-3fe4a6003995" (UID: "f1daafef-ea7f-42be-adf2-3fe4a6003995"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.013448 4959 scope.go:117] "RemoveContainer" containerID="cdef07dc561bd028518e79c39104ade93f6359395da1f024ed411caffa758ecc" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.022517 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1daafef-ea7f-42be-adf2-3fe4a6003995-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f1daafef-ea7f-42be-adf2-3fe4a6003995" (UID: "f1daafef-ea7f-42be-adf2-3fe4a6003995"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.025008 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1daafef-ea7f-42be-adf2-3fe4a6003995-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f1daafef-ea7f-42be-adf2-3fe4a6003995" (UID: "f1daafef-ea7f-42be-adf2-3fe4a6003995"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.034170 4959 scope.go:117] "RemoveContainer" containerID="e75946aa0c81914320f17158bf624ef5cdb2122addafae9457ba63d3cd2c8c88" Oct 07 14:06:16 crc kubenswrapper[4959]: E1007 14:06:16.034550 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e75946aa0c81914320f17158bf624ef5cdb2122addafae9457ba63d3cd2c8c88\": container with ID starting with e75946aa0c81914320f17158bf624ef5cdb2122addafae9457ba63d3cd2c8c88 not found: ID does not exist" containerID="e75946aa0c81914320f17158bf624ef5cdb2122addafae9457ba63d3cd2c8c88" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.034611 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e75946aa0c81914320f17158bf624ef5cdb2122addafae9457ba63d3cd2c8c88"} err="failed to get container status \"e75946aa0c81914320f17158bf624ef5cdb2122addafae9457ba63d3cd2c8c88\": rpc error: code = NotFound desc = could not find container \"e75946aa0c81914320f17158bf624ef5cdb2122addafae9457ba63d3cd2c8c88\": container with ID starting with e75946aa0c81914320f17158bf624ef5cdb2122addafae9457ba63d3cd2c8c88 not found: ID does not exist" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.034635 4959 scope.go:117] "RemoveContainer" containerID="cdef07dc561bd028518e79c39104ade93f6359395da1f024ed411caffa758ecc" Oct 07 14:06:16 crc kubenswrapper[4959]: E1007 14:06:16.034898 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdef07dc561bd028518e79c39104ade93f6359395da1f024ed411caffa758ecc\": container with ID starting with cdef07dc561bd028518e79c39104ade93f6359395da1f024ed411caffa758ecc not found: ID does not exist" containerID="cdef07dc561bd028518e79c39104ade93f6359395da1f024ed411caffa758ecc" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.034918 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdef07dc561bd028518e79c39104ade93f6359395da1f024ed411caffa758ecc"} err="failed to get container status \"cdef07dc561bd028518e79c39104ade93f6359395da1f024ed411caffa758ecc\": rpc error: code = NotFound desc = could not find container \"cdef07dc561bd028518e79c39104ade93f6359395da1f024ed411caffa758ecc\": container with ID starting with cdef07dc561bd028518e79c39104ade93f6359395da1f024ed411caffa758ecc not found: ID does not exist" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.034931 4959 scope.go:117] "RemoveContainer" containerID="e75946aa0c81914320f17158bf624ef5cdb2122addafae9457ba63d3cd2c8c88" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.035188 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e75946aa0c81914320f17158bf624ef5cdb2122addafae9457ba63d3cd2c8c88"} err="failed to get container status \"e75946aa0c81914320f17158bf624ef5cdb2122addafae9457ba63d3cd2c8c88\": rpc error: code = NotFound desc = could not find container \"e75946aa0c81914320f17158bf624ef5cdb2122addafae9457ba63d3cd2c8c88\": container with ID starting with e75946aa0c81914320f17158bf624ef5cdb2122addafae9457ba63d3cd2c8c88 not found: ID does not exist" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.035209 4959 scope.go:117] "RemoveContainer" containerID="cdef07dc561bd028518e79c39104ade93f6359395da1f024ed411caffa758ecc" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.035545 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdef07dc561bd028518e79c39104ade93f6359395da1f024ed411caffa758ecc"} err="failed to get container status \"cdef07dc561bd028518e79c39104ade93f6359395da1f024ed411caffa758ecc\": rpc error: code = NotFound desc = could not find container \"cdef07dc561bd028518e79c39104ade93f6359395da1f024ed411caffa758ecc\": container with ID starting with cdef07dc561bd028518e79c39104ade93f6359395da1f024ed411caffa758ecc not found: ID does not exist" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.065542 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1daafef-ea7f-42be-adf2-3fe4a6003995-logs\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.065584 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wpwc\" (UniqueName: \"kubernetes.io/projected/f1daafef-ea7f-42be-adf2-3fe4a6003995-kube-api-access-6wpwc\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.065600 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1daafef-ea7f-42be-adf2-3fe4a6003995-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.065614 4959 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1daafef-ea7f-42be-adf2-3fe4a6003995-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.065626 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1daafef-ea7f-42be-adf2-3fe4a6003995-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.065638 4959 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1daafef-ea7f-42be-adf2-3fe4a6003995-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.323508 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.345458 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.353422 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 07 14:06:16 crc kubenswrapper[4959]: E1007 14:06:16.353902 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1daafef-ea7f-42be-adf2-3fe4a6003995" containerName="nova-api-log" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.353924 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1daafef-ea7f-42be-adf2-3fe4a6003995" containerName="nova-api-log" Oct 07 14:06:16 crc kubenswrapper[4959]: E1007 14:06:16.353944 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adff668e-a674-4ad1-ae8f-a28b2f6bc9e3" containerName="nova-manage" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.353953 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="adff668e-a674-4ad1-ae8f-a28b2f6bc9e3" containerName="nova-manage" Oct 07 14:06:16 crc kubenswrapper[4959]: E1007 14:06:16.353964 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc43cc8-61cc-4102-aacc-356966a9b9a5" containerName="dnsmasq-dns" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.353971 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc43cc8-61cc-4102-aacc-356966a9b9a5" containerName="dnsmasq-dns" Oct 07 14:06:16 crc kubenswrapper[4959]: E1007 14:06:16.354004 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1daafef-ea7f-42be-adf2-3fe4a6003995" containerName="nova-api-api" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.354013 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1daafef-ea7f-42be-adf2-3fe4a6003995" containerName="nova-api-api" Oct 07 14:06:16 crc kubenswrapper[4959]: E1007 14:06:16.354024 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc43cc8-61cc-4102-aacc-356966a9b9a5" containerName="init" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.354032 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc43cc8-61cc-4102-aacc-356966a9b9a5" containerName="init" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.354255 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="adff668e-a674-4ad1-ae8f-a28b2f6bc9e3" containerName="nova-manage" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.354275 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1daafef-ea7f-42be-adf2-3fe4a6003995" containerName="nova-api-log" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.354289 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1daafef-ea7f-42be-adf2-3fe4a6003995" containerName="nova-api-api" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.354304 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fc43cc8-61cc-4102-aacc-356966a9b9a5" containerName="dnsmasq-dns" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.355443 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.357666 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.357670 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.358888 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.363347 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.404784 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/557f7b70-2476-481e-afaf-fc346d47b007-config-data\") pod \"nova-api-0\" (UID: \"557f7b70-2476-481e-afaf-fc346d47b007\") " pod="openstack/nova-api-0" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.404835 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/557f7b70-2476-481e-afaf-fc346d47b007-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"557f7b70-2476-481e-afaf-fc346d47b007\") " pod="openstack/nova-api-0" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.405059 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lkfd\" (UniqueName: \"kubernetes.io/projected/557f7b70-2476-481e-afaf-fc346d47b007-kube-api-access-2lkfd\") pod \"nova-api-0\" (UID: \"557f7b70-2476-481e-afaf-fc346d47b007\") " pod="openstack/nova-api-0" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.405209 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/557f7b70-2476-481e-afaf-fc346d47b007-public-tls-certs\") pod \"nova-api-0\" (UID: \"557f7b70-2476-481e-afaf-fc346d47b007\") " pod="openstack/nova-api-0" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.405243 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/557f7b70-2476-481e-afaf-fc346d47b007-internal-tls-certs\") pod \"nova-api-0\" (UID: \"557f7b70-2476-481e-afaf-fc346d47b007\") " pod="openstack/nova-api-0" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.405271 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/557f7b70-2476-481e-afaf-fc346d47b007-logs\") pod \"nova-api-0\" (UID: \"557f7b70-2476-481e-afaf-fc346d47b007\") " pod="openstack/nova-api-0" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.506502 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/557f7b70-2476-481e-afaf-fc346d47b007-internal-tls-certs\") pod \"nova-api-0\" (UID: \"557f7b70-2476-481e-afaf-fc346d47b007\") " pod="openstack/nova-api-0" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.506563 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/557f7b70-2476-481e-afaf-fc346d47b007-logs\") pod \"nova-api-0\" (UID: \"557f7b70-2476-481e-afaf-fc346d47b007\") " pod="openstack/nova-api-0" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.506603 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/557f7b70-2476-481e-afaf-fc346d47b007-config-data\") pod \"nova-api-0\" (UID: \"557f7b70-2476-481e-afaf-fc346d47b007\") " pod="openstack/nova-api-0" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.506630 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/557f7b70-2476-481e-afaf-fc346d47b007-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"557f7b70-2476-481e-afaf-fc346d47b007\") " pod="openstack/nova-api-0" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.506901 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lkfd\" (UniqueName: \"kubernetes.io/projected/557f7b70-2476-481e-afaf-fc346d47b007-kube-api-access-2lkfd\") pod \"nova-api-0\" (UID: \"557f7b70-2476-481e-afaf-fc346d47b007\") " pod="openstack/nova-api-0" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.506940 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/557f7b70-2476-481e-afaf-fc346d47b007-logs\") pod \"nova-api-0\" (UID: \"557f7b70-2476-481e-afaf-fc346d47b007\") " pod="openstack/nova-api-0" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.507035 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/557f7b70-2476-481e-afaf-fc346d47b007-public-tls-certs\") pod \"nova-api-0\" (UID: \"557f7b70-2476-481e-afaf-fc346d47b007\") " pod="openstack/nova-api-0" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.509787 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/557f7b70-2476-481e-afaf-fc346d47b007-public-tls-certs\") pod \"nova-api-0\" (UID: \"557f7b70-2476-481e-afaf-fc346d47b007\") " pod="openstack/nova-api-0" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.510245 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/557f7b70-2476-481e-afaf-fc346d47b007-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"557f7b70-2476-481e-afaf-fc346d47b007\") " pod="openstack/nova-api-0" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.510269 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/557f7b70-2476-481e-afaf-fc346d47b007-config-data\") pod \"nova-api-0\" (UID: \"557f7b70-2476-481e-afaf-fc346d47b007\") " pod="openstack/nova-api-0" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.510577 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/557f7b70-2476-481e-afaf-fc346d47b007-internal-tls-certs\") pod \"nova-api-0\" (UID: \"557f7b70-2476-481e-afaf-fc346d47b007\") " pod="openstack/nova-api-0" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.533399 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lkfd\" (UniqueName: \"kubernetes.io/projected/557f7b70-2476-481e-afaf-fc346d47b007-kube-api-access-2lkfd\") pod \"nova-api-0\" (UID: \"557f7b70-2476-481e-afaf-fc346d47b007\") " pod="openstack/nova-api-0" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.663465 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1daafef-ea7f-42be-adf2-3fe4a6003995" path="/var/lib/kubelet/pods/f1daafef-ea7f-42be-adf2-3fe4a6003995/volumes" Oct 07 14:06:16 crc kubenswrapper[4959]: I1007 14:06:16.719397 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 07 14:06:16 crc kubenswrapper[4959]: E1007 14:06:16.883462 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="48cf7552cac7401d7fb488ed8ed9ab98b4408a52de7d47873248702d94af9ad5" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 07 14:06:16 crc kubenswrapper[4959]: E1007 14:06:16.889363 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="48cf7552cac7401d7fb488ed8ed9ab98b4408a52de7d47873248702d94af9ad5" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 07 14:06:16 crc kubenswrapper[4959]: E1007 14:06:16.891750 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="48cf7552cac7401d7fb488ed8ed9ab98b4408a52de7d47873248702d94af9ad5" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 07 14:06:16 crc kubenswrapper[4959]: E1007 14:06:16.891782 4959 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="14f0652c-3a52-460c-9ebc-86eeba72e77a" containerName="nova-scheduler-scheduler" Oct 07 14:06:17 crc kubenswrapper[4959]: I1007 14:06:17.157561 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 07 14:06:18 crc kubenswrapper[4959]: I1007 14:06:18.016190 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"557f7b70-2476-481e-afaf-fc346d47b007","Type":"ContainerStarted","Data":"3615887c13e531e73136db7090798ee0c76932d3364b96841c3978994240e5e3"} Oct 07 14:06:18 crc kubenswrapper[4959]: I1007 14:06:18.016717 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"557f7b70-2476-481e-afaf-fc346d47b007","Type":"ContainerStarted","Data":"d354e03076a873eaa1ca438c00c6e3a0dff74a356caa0bace07812c0b488e5af"} Oct 07 14:06:18 crc kubenswrapper[4959]: I1007 14:06:18.016729 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"557f7b70-2476-481e-afaf-fc346d47b007","Type":"ContainerStarted","Data":"eaace1563b4d5ee79e24b46532fb92cc59b9f5db7f5e3e67945ea4b2e4fc971f"} Oct 07 14:06:18 crc kubenswrapper[4959]: I1007 14:06:18.036575 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.036554238 podStartE2EDuration="2.036554238s" podCreationTimestamp="2025-10-07 14:06:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:06:18.032424225 +0000 UTC m=+1260.115828550" watchObservedRunningTime="2025-10-07 14:06:18.036554238 +0000 UTC m=+1260.119958553" Oct 07 14:06:18 crc kubenswrapper[4959]: I1007 14:06:18.356825 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="0a9c0a78-b2ae-4edb-8b49-be96b1b0731f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.184:8775/\": read tcp 10.217.0.2:42960->10.217.0.184:8775: read: connection reset by peer" Oct 07 14:06:18 crc kubenswrapper[4959]: I1007 14:06:18.356851 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="0a9c0a78-b2ae-4edb-8b49-be96b1b0731f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.184:8775/\": read tcp 10.217.0.2:42944->10.217.0.184:8775: read: connection reset by peer" Oct 07 14:06:18 crc kubenswrapper[4959]: I1007 14:06:18.827818 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 14:06:18 crc kubenswrapper[4959]: I1007 14:06:18.963398 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-combined-ca-bundle\") pod \"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f\" (UID: \"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f\") " Oct 07 14:06:18 crc kubenswrapper[4959]: I1007 14:06:18.963530 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcj89\" (UniqueName: \"kubernetes.io/projected/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-kube-api-access-dcj89\") pod \"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f\" (UID: \"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f\") " Oct 07 14:06:18 crc kubenswrapper[4959]: I1007 14:06:18.963587 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-nova-metadata-tls-certs\") pod \"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f\" (UID: \"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f\") " Oct 07 14:06:18 crc kubenswrapper[4959]: I1007 14:06:18.963622 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-logs\") pod \"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f\" (UID: \"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f\") " Oct 07 14:06:18 crc kubenswrapper[4959]: I1007 14:06:18.963691 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-config-data\") pod \"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f\" (UID: \"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f\") " Oct 07 14:06:18 crc kubenswrapper[4959]: I1007 14:06:18.964051 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-logs" (OuterVolumeSpecName: "logs") pod "0a9c0a78-b2ae-4edb-8b49-be96b1b0731f" (UID: "0a9c0a78-b2ae-4edb-8b49-be96b1b0731f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:06:18 crc kubenswrapper[4959]: I1007 14:06:18.968046 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-kube-api-access-dcj89" (OuterVolumeSpecName: "kube-api-access-dcj89") pod "0a9c0a78-b2ae-4edb-8b49-be96b1b0731f" (UID: "0a9c0a78-b2ae-4edb-8b49-be96b1b0731f"). InnerVolumeSpecName "kube-api-access-dcj89". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:06:18 crc kubenswrapper[4959]: I1007 14:06:18.988786 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-config-data" (OuterVolumeSpecName: "config-data") pod "0a9c0a78-b2ae-4edb-8b49-be96b1b0731f" (UID: "0a9c0a78-b2ae-4edb-8b49-be96b1b0731f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:06:18 crc kubenswrapper[4959]: I1007 14:06:18.997515 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a9c0a78-b2ae-4edb-8b49-be96b1b0731f" (UID: "0a9c0a78-b2ae-4edb-8b49-be96b1b0731f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.015695 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "0a9c0a78-b2ae-4edb-8b49-be96b1b0731f" (UID: "0a9c0a78-b2ae-4edb-8b49-be96b1b0731f"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.026676 4959 generic.go:334] "Generic (PLEG): container finished" podID="0a9c0a78-b2ae-4edb-8b49-be96b1b0731f" containerID="8cd1268f6a16cc19070e1500a6303258aa9ca5e7bfa867d9350c4ac21ee4d585" exitCode=0 Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.026728 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f","Type":"ContainerDied","Data":"8cd1268f6a16cc19070e1500a6303258aa9ca5e7bfa867d9350c4ac21ee4d585"} Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.026774 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0a9c0a78-b2ae-4edb-8b49-be96b1b0731f","Type":"ContainerDied","Data":"e27732ede22f998af89667530ea54c3ca08cfbd63240c3ae61b6f2a23fec75aa"} Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.026791 4959 scope.go:117] "RemoveContainer" containerID="8cd1268f6a16cc19070e1500a6303258aa9ca5e7bfa867d9350c4ac21ee4d585" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.026808 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.067615 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.067652 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.067667 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcj89\" (UniqueName: \"kubernetes.io/projected/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-kube-api-access-dcj89\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.067680 4959 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.067691 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f-logs\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.074328 4959 scope.go:117] "RemoveContainer" containerID="c24cd83763ffe59dcb7a1d47b5f5811d39efff6aeb0689d5ea52a3e526d61ba4" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.088738 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.095648 4959 scope.go:117] "RemoveContainer" containerID="8cd1268f6a16cc19070e1500a6303258aa9ca5e7bfa867d9350c4ac21ee4d585" Oct 07 14:06:19 crc kubenswrapper[4959]: E1007 14:06:19.096183 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cd1268f6a16cc19070e1500a6303258aa9ca5e7bfa867d9350c4ac21ee4d585\": container with ID starting with 8cd1268f6a16cc19070e1500a6303258aa9ca5e7bfa867d9350c4ac21ee4d585 not found: ID does not exist" containerID="8cd1268f6a16cc19070e1500a6303258aa9ca5e7bfa867d9350c4ac21ee4d585" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.096251 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cd1268f6a16cc19070e1500a6303258aa9ca5e7bfa867d9350c4ac21ee4d585"} err="failed to get container status \"8cd1268f6a16cc19070e1500a6303258aa9ca5e7bfa867d9350c4ac21ee4d585\": rpc error: code = NotFound desc = could not find container \"8cd1268f6a16cc19070e1500a6303258aa9ca5e7bfa867d9350c4ac21ee4d585\": container with ID starting with 8cd1268f6a16cc19070e1500a6303258aa9ca5e7bfa867d9350c4ac21ee4d585 not found: ID does not exist" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.096297 4959 scope.go:117] "RemoveContainer" containerID="c24cd83763ffe59dcb7a1d47b5f5811d39efff6aeb0689d5ea52a3e526d61ba4" Oct 07 14:06:19 crc kubenswrapper[4959]: E1007 14:06:19.096693 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c24cd83763ffe59dcb7a1d47b5f5811d39efff6aeb0689d5ea52a3e526d61ba4\": container with ID starting with c24cd83763ffe59dcb7a1d47b5f5811d39efff6aeb0689d5ea52a3e526d61ba4 not found: ID does not exist" containerID="c24cd83763ffe59dcb7a1d47b5f5811d39efff6aeb0689d5ea52a3e526d61ba4" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.096730 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c24cd83763ffe59dcb7a1d47b5f5811d39efff6aeb0689d5ea52a3e526d61ba4"} err="failed to get container status \"c24cd83763ffe59dcb7a1d47b5f5811d39efff6aeb0689d5ea52a3e526d61ba4\": rpc error: code = NotFound desc = could not find container \"c24cd83763ffe59dcb7a1d47b5f5811d39efff6aeb0689d5ea52a3e526d61ba4\": container with ID starting with c24cd83763ffe59dcb7a1d47b5f5811d39efff6aeb0689d5ea52a3e526d61ba4 not found: ID does not exist" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.097818 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.121405 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 07 14:06:19 crc kubenswrapper[4959]: E1007 14:06:19.121806 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a9c0a78-b2ae-4edb-8b49-be96b1b0731f" containerName="nova-metadata-metadata" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.121826 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a9c0a78-b2ae-4edb-8b49-be96b1b0731f" containerName="nova-metadata-metadata" Oct 07 14:06:19 crc kubenswrapper[4959]: E1007 14:06:19.121844 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a9c0a78-b2ae-4edb-8b49-be96b1b0731f" containerName="nova-metadata-log" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.121853 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a9c0a78-b2ae-4edb-8b49-be96b1b0731f" containerName="nova-metadata-log" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.122066 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a9c0a78-b2ae-4edb-8b49-be96b1b0731f" containerName="nova-metadata-log" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.122099 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a9c0a78-b2ae-4edb-8b49-be96b1b0731f" containerName="nova-metadata-metadata" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.123167 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.131513 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.131811 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.137055 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.274760 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvnm5\" (UniqueName: \"kubernetes.io/projected/a046abc0-ec2e-42ac-ad49-e336824e9317-kube-api-access-hvnm5\") pod \"nova-metadata-0\" (UID: \"a046abc0-ec2e-42ac-ad49-e336824e9317\") " pod="openstack/nova-metadata-0" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.275262 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a046abc0-ec2e-42ac-ad49-e336824e9317-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a046abc0-ec2e-42ac-ad49-e336824e9317\") " pod="openstack/nova-metadata-0" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.275627 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a046abc0-ec2e-42ac-ad49-e336824e9317-logs\") pod \"nova-metadata-0\" (UID: \"a046abc0-ec2e-42ac-ad49-e336824e9317\") " pod="openstack/nova-metadata-0" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.275899 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a046abc0-ec2e-42ac-ad49-e336824e9317-config-data\") pod \"nova-metadata-0\" (UID: \"a046abc0-ec2e-42ac-ad49-e336824e9317\") " pod="openstack/nova-metadata-0" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.276186 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a046abc0-ec2e-42ac-ad49-e336824e9317-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a046abc0-ec2e-42ac-ad49-e336824e9317\") " pod="openstack/nova-metadata-0" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.378706 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvnm5\" (UniqueName: \"kubernetes.io/projected/a046abc0-ec2e-42ac-ad49-e336824e9317-kube-api-access-hvnm5\") pod \"nova-metadata-0\" (UID: \"a046abc0-ec2e-42ac-ad49-e336824e9317\") " pod="openstack/nova-metadata-0" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.378796 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a046abc0-ec2e-42ac-ad49-e336824e9317-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a046abc0-ec2e-42ac-ad49-e336824e9317\") " pod="openstack/nova-metadata-0" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.378839 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a046abc0-ec2e-42ac-ad49-e336824e9317-logs\") pod \"nova-metadata-0\" (UID: \"a046abc0-ec2e-42ac-ad49-e336824e9317\") " pod="openstack/nova-metadata-0" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.378895 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a046abc0-ec2e-42ac-ad49-e336824e9317-config-data\") pod \"nova-metadata-0\" (UID: \"a046abc0-ec2e-42ac-ad49-e336824e9317\") " pod="openstack/nova-metadata-0" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.378911 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a046abc0-ec2e-42ac-ad49-e336824e9317-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a046abc0-ec2e-42ac-ad49-e336824e9317\") " pod="openstack/nova-metadata-0" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.379898 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a046abc0-ec2e-42ac-ad49-e336824e9317-logs\") pod \"nova-metadata-0\" (UID: \"a046abc0-ec2e-42ac-ad49-e336824e9317\") " pod="openstack/nova-metadata-0" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.382203 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a046abc0-ec2e-42ac-ad49-e336824e9317-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a046abc0-ec2e-42ac-ad49-e336824e9317\") " pod="openstack/nova-metadata-0" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.382756 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a046abc0-ec2e-42ac-ad49-e336824e9317-config-data\") pod \"nova-metadata-0\" (UID: \"a046abc0-ec2e-42ac-ad49-e336824e9317\") " pod="openstack/nova-metadata-0" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.383869 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a046abc0-ec2e-42ac-ad49-e336824e9317-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a046abc0-ec2e-42ac-ad49-e336824e9317\") " pod="openstack/nova-metadata-0" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.395849 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvnm5\" (UniqueName: \"kubernetes.io/projected/a046abc0-ec2e-42ac-ad49-e336824e9317-kube-api-access-hvnm5\") pod \"nova-metadata-0\" (UID: \"a046abc0-ec2e-42ac-ad49-e336824e9317\") " pod="openstack/nova-metadata-0" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.478879 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 07 14:06:19 crc kubenswrapper[4959]: I1007 14:06:19.907700 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 07 14:06:20 crc kubenswrapper[4959]: I1007 14:06:20.036749 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a046abc0-ec2e-42ac-ad49-e336824e9317","Type":"ContainerStarted","Data":"f8c6440c013c5ee837858457bc918b3993c208b5ebffa354a1714bf65e90bd24"} Oct 07 14:06:20 crc kubenswrapper[4959]: I1007 14:06:20.673985 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a9c0a78-b2ae-4edb-8b49-be96b1b0731f" path="/var/lib/kubelet/pods/0a9c0a78-b2ae-4edb-8b49-be96b1b0731f/volumes" Oct 07 14:06:20 crc kubenswrapper[4959]: E1007 14:06:20.866717 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod14f0652c_3a52_460c_9ebc_86eeba72e77a.slice/crio-conmon-48cf7552cac7401d7fb488ed8ed9ab98b4408a52de7d47873248702d94af9ad5.scope\": RecentStats: unable to find data in memory cache]" Oct 07 14:06:21 crc kubenswrapper[4959]: I1007 14:06:21.052409 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a046abc0-ec2e-42ac-ad49-e336824e9317","Type":"ContainerStarted","Data":"a1280bf98239beaf2e37ec7fee5c562d280e63784aadda4b3eff0ffe23d02ee3"} Oct 07 14:06:21 crc kubenswrapper[4959]: I1007 14:06:21.052460 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a046abc0-ec2e-42ac-ad49-e336824e9317","Type":"ContainerStarted","Data":"5ee786d9237582cd9483cde7a6772ac58cb78bd6ec16ae873776b71c84740981"} Oct 07 14:06:21 crc kubenswrapper[4959]: I1007 14:06:21.056187 4959 generic.go:334] "Generic (PLEG): container finished" podID="14f0652c-3a52-460c-9ebc-86eeba72e77a" containerID="48cf7552cac7401d7fb488ed8ed9ab98b4408a52de7d47873248702d94af9ad5" exitCode=0 Oct 07 14:06:21 crc kubenswrapper[4959]: I1007 14:06:21.056221 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"14f0652c-3a52-460c-9ebc-86eeba72e77a","Type":"ContainerDied","Data":"48cf7552cac7401d7fb488ed8ed9ab98b4408a52de7d47873248702d94af9ad5"} Oct 07 14:06:21 crc kubenswrapper[4959]: I1007 14:06:21.082542 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.082519968 podStartE2EDuration="2.082519968s" podCreationTimestamp="2025-10-07 14:06:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:06:21.077305497 +0000 UTC m=+1263.160709832" watchObservedRunningTime="2025-10-07 14:06:21.082519968 +0000 UTC m=+1263.165924303" Oct 07 14:06:21 crc kubenswrapper[4959]: I1007 14:06:21.637447 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 14:06:21 crc kubenswrapper[4959]: I1007 14:06:21.735817 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14f0652c-3a52-460c-9ebc-86eeba72e77a-combined-ca-bundle\") pod \"14f0652c-3a52-460c-9ebc-86eeba72e77a\" (UID: \"14f0652c-3a52-460c-9ebc-86eeba72e77a\") " Oct 07 14:06:21 crc kubenswrapper[4959]: I1007 14:06:21.735882 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvls2\" (UniqueName: \"kubernetes.io/projected/14f0652c-3a52-460c-9ebc-86eeba72e77a-kube-api-access-kvls2\") pod \"14f0652c-3a52-460c-9ebc-86eeba72e77a\" (UID: \"14f0652c-3a52-460c-9ebc-86eeba72e77a\") " Oct 07 14:06:21 crc kubenswrapper[4959]: I1007 14:06:21.735959 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14f0652c-3a52-460c-9ebc-86eeba72e77a-config-data\") pod \"14f0652c-3a52-460c-9ebc-86eeba72e77a\" (UID: \"14f0652c-3a52-460c-9ebc-86eeba72e77a\") " Oct 07 14:06:21 crc kubenswrapper[4959]: I1007 14:06:21.743270 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14f0652c-3a52-460c-9ebc-86eeba72e77a-kube-api-access-kvls2" (OuterVolumeSpecName: "kube-api-access-kvls2") pod "14f0652c-3a52-460c-9ebc-86eeba72e77a" (UID: "14f0652c-3a52-460c-9ebc-86eeba72e77a"). InnerVolumeSpecName "kube-api-access-kvls2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:06:21 crc kubenswrapper[4959]: I1007 14:06:21.779450 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14f0652c-3a52-460c-9ebc-86eeba72e77a-config-data" (OuterVolumeSpecName: "config-data") pod "14f0652c-3a52-460c-9ebc-86eeba72e77a" (UID: "14f0652c-3a52-460c-9ebc-86eeba72e77a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:06:21 crc kubenswrapper[4959]: I1007 14:06:21.784907 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14f0652c-3a52-460c-9ebc-86eeba72e77a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14f0652c-3a52-460c-9ebc-86eeba72e77a" (UID: "14f0652c-3a52-460c-9ebc-86eeba72e77a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:06:21 crc kubenswrapper[4959]: I1007 14:06:21.837852 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14f0652c-3a52-460c-9ebc-86eeba72e77a-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:21 crc kubenswrapper[4959]: I1007 14:06:21.837884 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14f0652c-3a52-460c-9ebc-86eeba72e77a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:21 crc kubenswrapper[4959]: I1007 14:06:21.837897 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvls2\" (UniqueName: \"kubernetes.io/projected/14f0652c-3a52-460c-9ebc-86eeba72e77a-kube-api-access-kvls2\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:22 crc kubenswrapper[4959]: I1007 14:06:22.067796 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 14:06:22 crc kubenswrapper[4959]: I1007 14:06:22.067797 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"14f0652c-3a52-460c-9ebc-86eeba72e77a","Type":"ContainerDied","Data":"cd76603308801e42635139edd46ec5c0b186132797e4230ec2650b1e3d588c22"} Oct 07 14:06:22 crc kubenswrapper[4959]: I1007 14:06:22.067859 4959 scope.go:117] "RemoveContainer" containerID="48cf7552cac7401d7fb488ed8ed9ab98b4408a52de7d47873248702d94af9ad5" Oct 07 14:06:22 crc kubenswrapper[4959]: I1007 14:06:22.107169 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 14:06:22 crc kubenswrapper[4959]: I1007 14:06:22.125377 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 14:06:22 crc kubenswrapper[4959]: I1007 14:06:22.140960 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 14:06:22 crc kubenswrapper[4959]: E1007 14:06:22.141490 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f0652c-3a52-460c-9ebc-86eeba72e77a" containerName="nova-scheduler-scheduler" Oct 07 14:06:22 crc kubenswrapper[4959]: I1007 14:06:22.141512 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f0652c-3a52-460c-9ebc-86eeba72e77a" containerName="nova-scheduler-scheduler" Oct 07 14:06:22 crc kubenswrapper[4959]: I1007 14:06:22.141737 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f0652c-3a52-460c-9ebc-86eeba72e77a" containerName="nova-scheduler-scheduler" Oct 07 14:06:22 crc kubenswrapper[4959]: I1007 14:06:22.142587 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 14:06:22 crc kubenswrapper[4959]: I1007 14:06:22.144239 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 07 14:06:22 crc kubenswrapper[4959]: I1007 14:06:22.145103 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 14:06:22 crc kubenswrapper[4959]: I1007 14:06:22.245623 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4msct\" (UniqueName: \"kubernetes.io/projected/22fed6d1-09b4-4d80-8fde-d3759d994af1-kube-api-access-4msct\") pod \"nova-scheduler-0\" (UID: \"22fed6d1-09b4-4d80-8fde-d3759d994af1\") " pod="openstack/nova-scheduler-0" Oct 07 14:06:22 crc kubenswrapper[4959]: I1007 14:06:22.245686 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22fed6d1-09b4-4d80-8fde-d3759d994af1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"22fed6d1-09b4-4d80-8fde-d3759d994af1\") " pod="openstack/nova-scheduler-0" Oct 07 14:06:22 crc kubenswrapper[4959]: I1007 14:06:22.246103 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22fed6d1-09b4-4d80-8fde-d3759d994af1-config-data\") pod \"nova-scheduler-0\" (UID: \"22fed6d1-09b4-4d80-8fde-d3759d994af1\") " pod="openstack/nova-scheduler-0" Oct 07 14:06:22 crc kubenswrapper[4959]: I1007 14:06:22.348209 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4msct\" (UniqueName: \"kubernetes.io/projected/22fed6d1-09b4-4d80-8fde-d3759d994af1-kube-api-access-4msct\") pod \"nova-scheduler-0\" (UID: \"22fed6d1-09b4-4d80-8fde-d3759d994af1\") " pod="openstack/nova-scheduler-0" Oct 07 14:06:22 crc kubenswrapper[4959]: I1007 14:06:22.348260 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22fed6d1-09b4-4d80-8fde-d3759d994af1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"22fed6d1-09b4-4d80-8fde-d3759d994af1\") " pod="openstack/nova-scheduler-0" Oct 07 14:06:22 crc kubenswrapper[4959]: I1007 14:06:22.348293 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22fed6d1-09b4-4d80-8fde-d3759d994af1-config-data\") pod \"nova-scheduler-0\" (UID: \"22fed6d1-09b4-4d80-8fde-d3759d994af1\") " pod="openstack/nova-scheduler-0" Oct 07 14:06:22 crc kubenswrapper[4959]: I1007 14:06:22.352406 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22fed6d1-09b4-4d80-8fde-d3759d994af1-config-data\") pod \"nova-scheduler-0\" (UID: \"22fed6d1-09b4-4d80-8fde-d3759d994af1\") " pod="openstack/nova-scheduler-0" Oct 07 14:06:22 crc kubenswrapper[4959]: I1007 14:06:22.352771 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22fed6d1-09b4-4d80-8fde-d3759d994af1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"22fed6d1-09b4-4d80-8fde-d3759d994af1\") " pod="openstack/nova-scheduler-0" Oct 07 14:06:22 crc kubenswrapper[4959]: I1007 14:06:22.364606 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4msct\" (UniqueName: \"kubernetes.io/projected/22fed6d1-09b4-4d80-8fde-d3759d994af1-kube-api-access-4msct\") pod \"nova-scheduler-0\" (UID: \"22fed6d1-09b4-4d80-8fde-d3759d994af1\") " pod="openstack/nova-scheduler-0" Oct 07 14:06:22 crc kubenswrapper[4959]: I1007 14:06:22.462658 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 07 14:06:22 crc kubenswrapper[4959]: I1007 14:06:22.665707 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14f0652c-3a52-460c-9ebc-86eeba72e77a" path="/var/lib/kubelet/pods/14f0652c-3a52-460c-9ebc-86eeba72e77a/volumes" Oct 07 14:06:22 crc kubenswrapper[4959]: I1007 14:06:22.950906 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 07 14:06:23 crc kubenswrapper[4959]: I1007 14:06:23.099147 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"22fed6d1-09b4-4d80-8fde-d3759d994af1","Type":"ContainerStarted","Data":"7cd3efb8d0164b506c9688b0102e698017d26f38216102faf44232dc352142d3"} Oct 07 14:06:24 crc kubenswrapper[4959]: I1007 14:06:24.107724 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"22fed6d1-09b4-4d80-8fde-d3759d994af1","Type":"ContainerStarted","Data":"db4abcae98bb473f0a2a6d8bea584d2d06515a83ce5765babcbaff3c609a793d"} Oct 07 14:06:24 crc kubenswrapper[4959]: I1007 14:06:24.127932 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.127912294 podStartE2EDuration="2.127912294s" podCreationTimestamp="2025-10-07 14:06:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:06:24.121697198 +0000 UTC m=+1266.205101523" watchObservedRunningTime="2025-10-07 14:06:24.127912294 +0000 UTC m=+1266.211316639" Oct 07 14:06:24 crc kubenswrapper[4959]: I1007 14:06:24.480033 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 07 14:06:24 crc kubenswrapper[4959]: I1007 14:06:24.480093 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 07 14:06:26 crc kubenswrapper[4959]: I1007 14:06:26.719645 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 07 14:06:26 crc kubenswrapper[4959]: I1007 14:06:26.720012 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 07 14:06:27 crc kubenswrapper[4959]: I1007 14:06:27.462849 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 07 14:06:27 crc kubenswrapper[4959]: I1007 14:06:27.733294 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="557f7b70-2476-481e-afaf-fc346d47b007" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.193:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 07 14:06:27 crc kubenswrapper[4959]: I1007 14:06:27.733320 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="557f7b70-2476-481e-afaf-fc346d47b007" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.193:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 07 14:06:29 crc kubenswrapper[4959]: I1007 14:06:29.479740 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 07 14:06:29 crc kubenswrapper[4959]: I1007 14:06:29.480077 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 07 14:06:30 crc kubenswrapper[4959]: I1007 14:06:30.488357 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a046abc0-ec2e-42ac-ad49-e336824e9317" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 07 14:06:30 crc kubenswrapper[4959]: I1007 14:06:30.488414 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a046abc0-ec2e-42ac-ad49-e336824e9317" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 07 14:06:32 crc kubenswrapper[4959]: I1007 14:06:32.463592 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 07 14:06:32 crc kubenswrapper[4959]: I1007 14:06:32.487416 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 07 14:06:33 crc kubenswrapper[4959]: I1007 14:06:33.217423 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 07 14:06:34 crc kubenswrapper[4959]: I1007 14:06:34.195438 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 07 14:06:36 crc kubenswrapper[4959]: I1007 14:06:36.730943 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 07 14:06:36 crc kubenswrapper[4959]: I1007 14:06:36.731561 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 07 14:06:36 crc kubenswrapper[4959]: I1007 14:06:36.734007 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 07 14:06:36 crc kubenswrapper[4959]: I1007 14:06:36.739275 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 07 14:06:37 crc kubenswrapper[4959]: I1007 14:06:37.215023 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 07 14:06:37 crc kubenswrapper[4959]: I1007 14:06:37.228357 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 07 14:06:39 crc kubenswrapper[4959]: I1007 14:06:39.484509 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 07 14:06:39 crc kubenswrapper[4959]: I1007 14:06:39.484863 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 07 14:06:39 crc kubenswrapper[4959]: I1007 14:06:39.495275 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 07 14:06:39 crc kubenswrapper[4959]: I1007 14:06:39.495324 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 07 14:06:47 crc kubenswrapper[4959]: I1007 14:06:47.876643 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 14:06:49 crc kubenswrapper[4959]: I1007 14:06:49.221358 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 14:06:50 crc kubenswrapper[4959]: I1007 14:06:50.629940 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:06:50 crc kubenswrapper[4959]: I1007 14:06:50.630043 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:06:52 crc kubenswrapper[4959]: I1007 14:06:52.301672 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="90af7ce4-adf2-4722-a726-cc4d7d29056f" containerName="rabbitmq" containerID="cri-o://0a122a18fbfa2dec9e05dd83c778908878ede8e5d5c924fd00f613322872d7e8" gracePeriod=604796 Oct 07 14:06:53 crc kubenswrapper[4959]: I1007 14:06:53.378214 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="e206a4fa-c178-4f85-be59-46edf7e78833" containerName="rabbitmq" containerID="cri-o://ba386e7a023b2ed0780a071418693e842c7589ef16170ddd067523a4be613e71" gracePeriod=604796 Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.628998 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="e206a4fa-c178-4f85-be59-46edf7e78833" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.873172 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.895447 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/90af7ce4-adf2-4722-a726-cc4d7d29056f-rabbitmq-tls\") pod \"90af7ce4-adf2-4722-a726-cc4d7d29056f\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.895489 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/90af7ce4-adf2-4722-a726-cc4d7d29056f-rabbitmq-confd\") pod \"90af7ce4-adf2-4722-a726-cc4d7d29056f\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.895536 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/90af7ce4-adf2-4722-a726-cc4d7d29056f-plugins-conf\") pod \"90af7ce4-adf2-4722-a726-cc4d7d29056f\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.895563 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/90af7ce4-adf2-4722-a726-cc4d7d29056f-rabbitmq-plugins\") pod \"90af7ce4-adf2-4722-a726-cc4d7d29056f\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.895610 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/90af7ce4-adf2-4722-a726-cc4d7d29056f-config-data\") pod \"90af7ce4-adf2-4722-a726-cc4d7d29056f\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.895637 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"90af7ce4-adf2-4722-a726-cc4d7d29056f\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.895679 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tw6fp\" (UniqueName: \"kubernetes.io/projected/90af7ce4-adf2-4722-a726-cc4d7d29056f-kube-api-access-tw6fp\") pod \"90af7ce4-adf2-4722-a726-cc4d7d29056f\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.895704 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/90af7ce4-adf2-4722-a726-cc4d7d29056f-pod-info\") pod \"90af7ce4-adf2-4722-a726-cc4d7d29056f\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.895725 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/90af7ce4-adf2-4722-a726-cc4d7d29056f-server-conf\") pod \"90af7ce4-adf2-4722-a726-cc4d7d29056f\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.895833 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/90af7ce4-adf2-4722-a726-cc4d7d29056f-rabbitmq-erlang-cookie\") pod \"90af7ce4-adf2-4722-a726-cc4d7d29056f\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.895916 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/90af7ce4-adf2-4722-a726-cc4d7d29056f-erlang-cookie-secret\") pod \"90af7ce4-adf2-4722-a726-cc4d7d29056f\" (UID: \"90af7ce4-adf2-4722-a726-cc4d7d29056f\") " Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.900449 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90af7ce4-adf2-4722-a726-cc4d7d29056f-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "90af7ce4-adf2-4722-a726-cc4d7d29056f" (UID: "90af7ce4-adf2-4722-a726-cc4d7d29056f"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.901654 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90af7ce4-adf2-4722-a726-cc4d7d29056f-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "90af7ce4-adf2-4722-a726-cc4d7d29056f" (UID: "90af7ce4-adf2-4722-a726-cc4d7d29056f"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.901722 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90af7ce4-adf2-4722-a726-cc4d7d29056f-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "90af7ce4-adf2-4722-a726-cc4d7d29056f" (UID: "90af7ce4-adf2-4722-a726-cc4d7d29056f"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.903349 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90af7ce4-adf2-4722-a726-cc4d7d29056f-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "90af7ce4-adf2-4722-a726-cc4d7d29056f" (UID: "90af7ce4-adf2-4722-a726-cc4d7d29056f"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.903561 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90af7ce4-adf2-4722-a726-cc4d7d29056f-kube-api-access-tw6fp" (OuterVolumeSpecName: "kube-api-access-tw6fp") pod "90af7ce4-adf2-4722-a726-cc4d7d29056f" (UID: "90af7ce4-adf2-4722-a726-cc4d7d29056f"). InnerVolumeSpecName "kube-api-access-tw6fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.905526 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "persistence") pod "90af7ce4-adf2-4722-a726-cc4d7d29056f" (UID: "90af7ce4-adf2-4722-a726-cc4d7d29056f"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.908768 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/90af7ce4-adf2-4722-a726-cc4d7d29056f-pod-info" (OuterVolumeSpecName: "pod-info") pod "90af7ce4-adf2-4722-a726-cc4d7d29056f" (UID: "90af7ce4-adf2-4722-a726-cc4d7d29056f"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.910825 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90af7ce4-adf2-4722-a726-cc4d7d29056f-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "90af7ce4-adf2-4722-a726-cc4d7d29056f" (UID: "90af7ce4-adf2-4722-a726-cc4d7d29056f"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.943355 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90af7ce4-adf2-4722-a726-cc4d7d29056f-config-data" (OuterVolumeSpecName: "config-data") pod "90af7ce4-adf2-4722-a726-cc4d7d29056f" (UID: "90af7ce4-adf2-4722-a726-cc4d7d29056f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.980677 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90af7ce4-adf2-4722-a726-cc4d7d29056f-server-conf" (OuterVolumeSpecName: "server-conf") pod "90af7ce4-adf2-4722-a726-cc4d7d29056f" (UID: "90af7ce4-adf2-4722-a726-cc4d7d29056f"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.997535 4959 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/90af7ce4-adf2-4722-a726-cc4d7d29056f-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.997568 4959 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/90af7ce4-adf2-4722-a726-cc4d7d29056f-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.997577 4959 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/90af7ce4-adf2-4722-a726-cc4d7d29056f-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.997585 4959 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/90af7ce4-adf2-4722-a726-cc4d7d29056f-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.997593 4959 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/90af7ce4-adf2-4722-a726-cc4d7d29056f-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.997602 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/90af7ce4-adf2-4722-a726-cc4d7d29056f-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.997627 4959 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.997636 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tw6fp\" (UniqueName: \"kubernetes.io/projected/90af7ce4-adf2-4722-a726-cc4d7d29056f-kube-api-access-tw6fp\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.997644 4959 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/90af7ce4-adf2-4722-a726-cc4d7d29056f-pod-info\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:58 crc kubenswrapper[4959]: I1007 14:06:58.997652 4959 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/90af7ce4-adf2-4722-a726-cc4d7d29056f-server-conf\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.022671 4959 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.068613 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90af7ce4-adf2-4722-a726-cc4d7d29056f-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "90af7ce4-adf2-4722-a726-cc4d7d29056f" (UID: "90af7ce4-adf2-4722-a726-cc4d7d29056f"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.098723 4959 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.098753 4959 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/90af7ce4-adf2-4722-a726-cc4d7d29056f-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.455955 4959 generic.go:334] "Generic (PLEG): container finished" podID="90af7ce4-adf2-4722-a726-cc4d7d29056f" containerID="0a122a18fbfa2dec9e05dd83c778908878ede8e5d5c924fd00f613322872d7e8" exitCode=0 Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.456004 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"90af7ce4-adf2-4722-a726-cc4d7d29056f","Type":"ContainerDied","Data":"0a122a18fbfa2dec9e05dd83c778908878ede8e5d5c924fd00f613322872d7e8"} Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.456015 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.456034 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"90af7ce4-adf2-4722-a726-cc4d7d29056f","Type":"ContainerDied","Data":"d27b48c1a514c89af66f80c59ec74ff8ac32d2fe06105b31c679db28fabf3a52"} Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.456053 4959 scope.go:117] "RemoveContainer" containerID="0a122a18fbfa2dec9e05dd83c778908878ede8e5d5c924fd00f613322872d7e8" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.473381 4959 scope.go:117] "RemoveContainer" containerID="b6140a1530482e80a1f682355efe01fbc1dd47f08e5fd863d3a4a6fcf5355ee8" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.492929 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.500655 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.528846 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 14:06:59 crc kubenswrapper[4959]: E1007 14:06:59.529299 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90af7ce4-adf2-4722-a726-cc4d7d29056f" containerName="rabbitmq" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.529321 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="90af7ce4-adf2-4722-a726-cc4d7d29056f" containerName="rabbitmq" Oct 07 14:06:59 crc kubenswrapper[4959]: E1007 14:06:59.529348 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90af7ce4-adf2-4722-a726-cc4d7d29056f" containerName="setup-container" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.529357 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="90af7ce4-adf2-4722-a726-cc4d7d29056f" containerName="setup-container" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.529591 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="90af7ce4-adf2-4722-a726-cc4d7d29056f" containerName="rabbitmq" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.530870 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.536658 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.536711 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.536854 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.536988 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.539427 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.539784 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.539933 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-8jwt2" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.543821 4959 scope.go:117] "RemoveContainer" containerID="0a122a18fbfa2dec9e05dd83c778908878ede8e5d5c924fd00f613322872d7e8" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.544390 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 14:06:59 crc kubenswrapper[4959]: E1007 14:06:59.544954 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a122a18fbfa2dec9e05dd83c778908878ede8e5d5c924fd00f613322872d7e8\": container with ID starting with 0a122a18fbfa2dec9e05dd83c778908878ede8e5d5c924fd00f613322872d7e8 not found: ID does not exist" containerID="0a122a18fbfa2dec9e05dd83c778908878ede8e5d5c924fd00f613322872d7e8" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.544998 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a122a18fbfa2dec9e05dd83c778908878ede8e5d5c924fd00f613322872d7e8"} err="failed to get container status \"0a122a18fbfa2dec9e05dd83c778908878ede8e5d5c924fd00f613322872d7e8\": rpc error: code = NotFound desc = could not find container \"0a122a18fbfa2dec9e05dd83c778908878ede8e5d5c924fd00f613322872d7e8\": container with ID starting with 0a122a18fbfa2dec9e05dd83c778908878ede8e5d5c924fd00f613322872d7e8 not found: ID does not exist" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.545020 4959 scope.go:117] "RemoveContainer" containerID="b6140a1530482e80a1f682355efe01fbc1dd47f08e5fd863d3a4a6fcf5355ee8" Oct 07 14:06:59 crc kubenswrapper[4959]: E1007 14:06:59.550481 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6140a1530482e80a1f682355efe01fbc1dd47f08e5fd863d3a4a6fcf5355ee8\": container with ID starting with b6140a1530482e80a1f682355efe01fbc1dd47f08e5fd863d3a4a6fcf5355ee8 not found: ID does not exist" containerID="b6140a1530482e80a1f682355efe01fbc1dd47f08e5fd863d3a4a6fcf5355ee8" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.550535 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6140a1530482e80a1f682355efe01fbc1dd47f08e5fd863d3a4a6fcf5355ee8"} err="failed to get container status \"b6140a1530482e80a1f682355efe01fbc1dd47f08e5fd863d3a4a6fcf5355ee8\": rpc error: code = NotFound desc = could not find container \"b6140a1530482e80a1f682355efe01fbc1dd47f08e5fd863d3a4a6fcf5355ee8\": container with ID starting with b6140a1530482e80a1f682355efe01fbc1dd47f08e5fd863d3a4a6fcf5355ee8 not found: ID does not exist" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.709477 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzwt8\" (UniqueName: \"kubernetes.io/projected/439982d5-b039-45c1-9300-f39ce049cf33-kube-api-access-vzwt8\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.709883 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/439982d5-b039-45c1-9300-f39ce049cf33-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.709912 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/439982d5-b039-45c1-9300-f39ce049cf33-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.709931 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/439982d5-b039-45c1-9300-f39ce049cf33-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.709945 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/439982d5-b039-45c1-9300-f39ce049cf33-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.709985 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/439982d5-b039-45c1-9300-f39ce049cf33-server-conf\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.710008 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/439982d5-b039-45c1-9300-f39ce049cf33-config-data\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.710043 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/439982d5-b039-45c1-9300-f39ce049cf33-pod-info\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.710062 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/439982d5-b039-45c1-9300-f39ce049cf33-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.710083 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/439982d5-b039-45c1-9300-f39ce049cf33-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.710100 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.811766 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/439982d5-b039-45c1-9300-f39ce049cf33-config-data\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.811833 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/439982d5-b039-45c1-9300-f39ce049cf33-pod-info\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.811887 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/439982d5-b039-45c1-9300-f39ce049cf33-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.811973 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/439982d5-b039-45c1-9300-f39ce049cf33-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.812005 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.812075 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzwt8\" (UniqueName: \"kubernetes.io/projected/439982d5-b039-45c1-9300-f39ce049cf33-kube-api-access-vzwt8\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.812156 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/439982d5-b039-45c1-9300-f39ce049cf33-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.812181 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/439982d5-b039-45c1-9300-f39ce049cf33-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.812204 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/439982d5-b039-45c1-9300-f39ce049cf33-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.812218 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/439982d5-b039-45c1-9300-f39ce049cf33-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.812264 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/439982d5-b039-45c1-9300-f39ce049cf33-server-conf\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.814197 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/439982d5-b039-45c1-9300-f39ce049cf33-server-conf\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.814934 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/439982d5-b039-45c1-9300-f39ce049cf33-config-data\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.815049 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/439982d5-b039-45c1-9300-f39ce049cf33-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.815487 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.817010 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/439982d5-b039-45c1-9300-f39ce049cf33-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.817020 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/439982d5-b039-45c1-9300-f39ce049cf33-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.820368 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/439982d5-b039-45c1-9300-f39ce049cf33-pod-info\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.820416 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/439982d5-b039-45c1-9300-f39ce049cf33-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.820845 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/439982d5-b039-45c1-9300-f39ce049cf33-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.831548 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/439982d5-b039-45c1-9300-f39ce049cf33-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.834154 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzwt8\" (UniqueName: \"kubernetes.io/projected/439982d5-b039-45c1-9300-f39ce049cf33-kube-api-access-vzwt8\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.853845 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"439982d5-b039-45c1-9300-f39ce049cf33\") " pod="openstack/rabbitmq-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.937701 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:06:59 crc kubenswrapper[4959]: I1007 14:06:59.977383 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.014646 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e206a4fa-c178-4f85-be59-46edf7e78833-server-conf\") pod \"e206a4fa-c178-4f85-be59-46edf7e78833\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.014958 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e206a4fa-c178-4f85-be59-46edf7e78833-config-data\") pod \"e206a4fa-c178-4f85-be59-46edf7e78833\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.015013 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"e206a4fa-c178-4f85-be59-46edf7e78833\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.015035 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e206a4fa-c178-4f85-be59-46edf7e78833-rabbitmq-plugins\") pod \"e206a4fa-c178-4f85-be59-46edf7e78833\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.015075 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e206a4fa-c178-4f85-be59-46edf7e78833-rabbitmq-confd\") pod \"e206a4fa-c178-4f85-be59-46edf7e78833\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.015091 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e206a4fa-c178-4f85-be59-46edf7e78833-pod-info\") pod \"e206a4fa-c178-4f85-be59-46edf7e78833\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.015145 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e206a4fa-c178-4f85-be59-46edf7e78833-plugins-conf\") pod \"e206a4fa-c178-4f85-be59-46edf7e78833\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.015168 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbdmf\" (UniqueName: \"kubernetes.io/projected/e206a4fa-c178-4f85-be59-46edf7e78833-kube-api-access-fbdmf\") pod \"e206a4fa-c178-4f85-be59-46edf7e78833\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.015205 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e206a4fa-c178-4f85-be59-46edf7e78833-erlang-cookie-secret\") pod \"e206a4fa-c178-4f85-be59-46edf7e78833\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.015227 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e206a4fa-c178-4f85-be59-46edf7e78833-rabbitmq-erlang-cookie\") pod \"e206a4fa-c178-4f85-be59-46edf7e78833\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.015258 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e206a4fa-c178-4f85-be59-46edf7e78833-rabbitmq-tls\") pod \"e206a4fa-c178-4f85-be59-46edf7e78833\" (UID: \"e206a4fa-c178-4f85-be59-46edf7e78833\") " Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.015546 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e206a4fa-c178-4f85-be59-46edf7e78833-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "e206a4fa-c178-4f85-be59-46edf7e78833" (UID: "e206a4fa-c178-4f85-be59-46edf7e78833"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.015880 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e206a4fa-c178-4f85-be59-46edf7e78833-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "e206a4fa-c178-4f85-be59-46edf7e78833" (UID: "e206a4fa-c178-4f85-be59-46edf7e78833"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.016038 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e206a4fa-c178-4f85-be59-46edf7e78833-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "e206a4fa-c178-4f85-be59-46edf7e78833" (UID: "e206a4fa-c178-4f85-be59-46edf7e78833"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.016651 4959 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e206a4fa-c178-4f85-be59-46edf7e78833-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.016670 4959 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e206a4fa-c178-4f85-be59-46edf7e78833-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.016679 4959 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e206a4fa-c178-4f85-be59-46edf7e78833-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.019206 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/e206a4fa-c178-4f85-be59-46edf7e78833-pod-info" (OuterVolumeSpecName: "pod-info") pod "e206a4fa-c178-4f85-be59-46edf7e78833" (UID: "e206a4fa-c178-4f85-be59-46edf7e78833"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.020132 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "e206a4fa-c178-4f85-be59-46edf7e78833" (UID: "e206a4fa-c178-4f85-be59-46edf7e78833"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.027947 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e206a4fa-c178-4f85-be59-46edf7e78833-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "e206a4fa-c178-4f85-be59-46edf7e78833" (UID: "e206a4fa-c178-4f85-be59-46edf7e78833"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.028305 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e206a4fa-c178-4f85-be59-46edf7e78833-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "e206a4fa-c178-4f85-be59-46edf7e78833" (UID: "e206a4fa-c178-4f85-be59-46edf7e78833"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.028422 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e206a4fa-c178-4f85-be59-46edf7e78833-kube-api-access-fbdmf" (OuterVolumeSpecName: "kube-api-access-fbdmf") pod "e206a4fa-c178-4f85-be59-46edf7e78833" (UID: "e206a4fa-c178-4f85-be59-46edf7e78833"). InnerVolumeSpecName "kube-api-access-fbdmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.054664 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e206a4fa-c178-4f85-be59-46edf7e78833-config-data" (OuterVolumeSpecName: "config-data") pod "e206a4fa-c178-4f85-be59-46edf7e78833" (UID: "e206a4fa-c178-4f85-be59-46edf7e78833"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.081962 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e206a4fa-c178-4f85-be59-46edf7e78833-server-conf" (OuterVolumeSpecName: "server-conf") pod "e206a4fa-c178-4f85-be59-46edf7e78833" (UID: "e206a4fa-c178-4f85-be59-46edf7e78833"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.120259 4959 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.120292 4959 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e206a4fa-c178-4f85-be59-46edf7e78833-pod-info\") on node \"crc\" DevicePath \"\"" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.120302 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbdmf\" (UniqueName: \"kubernetes.io/projected/e206a4fa-c178-4f85-be59-46edf7e78833-kube-api-access-fbdmf\") on node \"crc\" DevicePath \"\"" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.120312 4959 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e206a4fa-c178-4f85-be59-46edf7e78833-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.120320 4959 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e206a4fa-c178-4f85-be59-46edf7e78833-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.120328 4959 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e206a4fa-c178-4f85-be59-46edf7e78833-server-conf\") on node \"crc\" DevicePath \"\"" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.120336 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e206a4fa-c178-4f85-be59-46edf7e78833-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.152176 4959 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.173236 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e206a4fa-c178-4f85-be59-46edf7e78833-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "e206a4fa-c178-4f85-be59-46edf7e78833" (UID: "e206a4fa-c178-4f85-be59-46edf7e78833"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.222583 4959 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.222626 4959 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e206a4fa-c178-4f85-be59-46edf7e78833-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.494602 4959 generic.go:334] "Generic (PLEG): container finished" podID="e206a4fa-c178-4f85-be59-46edf7e78833" containerID="ba386e7a023b2ed0780a071418693e842c7589ef16170ddd067523a4be613e71" exitCode=0 Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.494691 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.494699 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e206a4fa-c178-4f85-be59-46edf7e78833","Type":"ContainerDied","Data":"ba386e7a023b2ed0780a071418693e842c7589ef16170ddd067523a4be613e71"} Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.495152 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e206a4fa-c178-4f85-be59-46edf7e78833","Type":"ContainerDied","Data":"4c72abc157662698c847fb5ffdb71559f582932a858fb0d4e6cbe9c90d872ed4"} Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.495171 4959 scope.go:117] "RemoveContainer" containerID="ba386e7a023b2ed0780a071418693e842c7589ef16170ddd067523a4be613e71" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.524606 4959 scope.go:117] "RemoveContainer" containerID="a9137451d6a4f2897ca96d09e23e311123bffd7256a6e6f4b2714386a1d4c45c" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.527683 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 07 14:07:00 crc kubenswrapper[4959]: W1007 14:07:00.537936 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod439982d5_b039_45c1_9300_f39ce049cf33.slice/crio-739c6f0195e90152edc70dea4fdcbb93cf416f0187ffa45baefdf63a6b14f2ac WatchSource:0}: Error finding container 739c6f0195e90152edc70dea4fdcbb93cf416f0187ffa45baefdf63a6b14f2ac: Status 404 returned error can't find the container with id 739c6f0195e90152edc70dea4fdcbb93cf416f0187ffa45baefdf63a6b14f2ac Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.626342 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.640397 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.642984 4959 scope.go:117] "RemoveContainer" containerID="ba386e7a023b2ed0780a071418693e842c7589ef16170ddd067523a4be613e71" Oct 07 14:07:00 crc kubenswrapper[4959]: E1007 14:07:00.643617 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba386e7a023b2ed0780a071418693e842c7589ef16170ddd067523a4be613e71\": container with ID starting with ba386e7a023b2ed0780a071418693e842c7589ef16170ddd067523a4be613e71 not found: ID does not exist" containerID="ba386e7a023b2ed0780a071418693e842c7589ef16170ddd067523a4be613e71" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.643667 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba386e7a023b2ed0780a071418693e842c7589ef16170ddd067523a4be613e71"} err="failed to get container status \"ba386e7a023b2ed0780a071418693e842c7589ef16170ddd067523a4be613e71\": rpc error: code = NotFound desc = could not find container \"ba386e7a023b2ed0780a071418693e842c7589ef16170ddd067523a4be613e71\": container with ID starting with ba386e7a023b2ed0780a071418693e842c7589ef16170ddd067523a4be613e71 not found: ID does not exist" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.643706 4959 scope.go:117] "RemoveContainer" containerID="a9137451d6a4f2897ca96d09e23e311123bffd7256a6e6f4b2714386a1d4c45c" Oct 07 14:07:00 crc kubenswrapper[4959]: E1007 14:07:00.646363 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9137451d6a4f2897ca96d09e23e311123bffd7256a6e6f4b2714386a1d4c45c\": container with ID starting with a9137451d6a4f2897ca96d09e23e311123bffd7256a6e6f4b2714386a1d4c45c not found: ID does not exist" containerID="a9137451d6a4f2897ca96d09e23e311123bffd7256a6e6f4b2714386a1d4c45c" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.646423 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9137451d6a4f2897ca96d09e23e311123bffd7256a6e6f4b2714386a1d4c45c"} err="failed to get container status \"a9137451d6a4f2897ca96d09e23e311123bffd7256a6e6f4b2714386a1d4c45c\": rpc error: code = NotFound desc = could not find container \"a9137451d6a4f2897ca96d09e23e311123bffd7256a6e6f4b2714386a1d4c45c\": container with ID starting with a9137451d6a4f2897ca96d09e23e311123bffd7256a6e6f4b2714386a1d4c45c not found: ID does not exist" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.670400 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90af7ce4-adf2-4722-a726-cc4d7d29056f" path="/var/lib/kubelet/pods/90af7ce4-adf2-4722-a726-cc4d7d29056f/volumes" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.671296 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e206a4fa-c178-4f85-be59-46edf7e78833" path="/var/lib/kubelet/pods/e206a4fa-c178-4f85-be59-46edf7e78833/volumes" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.671864 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 14:07:00 crc kubenswrapper[4959]: E1007 14:07:00.672181 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e206a4fa-c178-4f85-be59-46edf7e78833" containerName="setup-container" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.672199 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e206a4fa-c178-4f85-be59-46edf7e78833" containerName="setup-container" Oct 07 14:07:00 crc kubenswrapper[4959]: E1007 14:07:00.672229 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e206a4fa-c178-4f85-be59-46edf7e78833" containerName="rabbitmq" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.672236 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e206a4fa-c178-4f85-be59-46edf7e78833" containerName="rabbitmq" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.672407 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="e206a4fa-c178-4f85-be59-46edf7e78833" containerName="rabbitmq" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.673309 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.675720 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.675804 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.675901 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.675955 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.676056 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.676083 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-xr4q8" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.678789 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.679480 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.729952 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0265cd4e-529d-40a6-a14c-6f39d6068633-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.730013 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0265cd4e-529d-40a6-a14c-6f39d6068633-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.730069 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0265cd4e-529d-40a6-a14c-6f39d6068633-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.730156 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0265cd4e-529d-40a6-a14c-6f39d6068633-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.730192 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0265cd4e-529d-40a6-a14c-6f39d6068633-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.730219 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0265cd4e-529d-40a6-a14c-6f39d6068633-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.730255 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.730279 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0265cd4e-529d-40a6-a14c-6f39d6068633-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.730304 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0265cd4e-529d-40a6-a14c-6f39d6068633-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.730383 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbqbx\" (UniqueName: \"kubernetes.io/projected/0265cd4e-529d-40a6-a14c-6f39d6068633-kube-api-access-pbqbx\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.730418 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0265cd4e-529d-40a6-a14c-6f39d6068633-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.831720 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0265cd4e-529d-40a6-a14c-6f39d6068633-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.831783 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0265cd4e-529d-40a6-a14c-6f39d6068633-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.831813 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0265cd4e-529d-40a6-a14c-6f39d6068633-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.831850 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0265cd4e-529d-40a6-a14c-6f39d6068633-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.831894 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0265cd4e-529d-40a6-a14c-6f39d6068633-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.831927 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0265cd4e-529d-40a6-a14c-6f39d6068633-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.831951 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0265cd4e-529d-40a6-a14c-6f39d6068633-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.831982 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.832005 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0265cd4e-529d-40a6-a14c-6f39d6068633-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.832031 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0265cd4e-529d-40a6-a14c-6f39d6068633-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.832127 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbqbx\" (UniqueName: \"kubernetes.io/projected/0265cd4e-529d-40a6-a14c-6f39d6068633-kube-api-access-pbqbx\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.832271 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0265cd4e-529d-40a6-a14c-6f39d6068633-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.832588 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0265cd4e-529d-40a6-a14c-6f39d6068633-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.832815 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.833387 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0265cd4e-529d-40a6-a14c-6f39d6068633-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.833542 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0265cd4e-529d-40a6-a14c-6f39d6068633-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.834211 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0265cd4e-529d-40a6-a14c-6f39d6068633-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.836503 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0265cd4e-529d-40a6-a14c-6f39d6068633-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.838875 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0265cd4e-529d-40a6-a14c-6f39d6068633-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.841507 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0265cd4e-529d-40a6-a14c-6f39d6068633-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.841570 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0265cd4e-529d-40a6-a14c-6f39d6068633-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.857506 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbqbx\" (UniqueName: \"kubernetes.io/projected/0265cd4e-529d-40a6-a14c-6f39d6068633-kube-api-access-pbqbx\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:00 crc kubenswrapper[4959]: I1007 14:07:00.873888 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0265cd4e-529d-40a6-a14c-6f39d6068633\") " pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:01 crc kubenswrapper[4959]: I1007 14:07:01.024416 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:01 crc kubenswrapper[4959]: I1007 14:07:01.484911 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 07 14:07:01 crc kubenswrapper[4959]: W1007 14:07:01.490083 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0265cd4e_529d_40a6_a14c_6f39d6068633.slice/crio-4d3fc474e2852c473505048a03f73d283a85e3e1e3f108988bdce1c714472460 WatchSource:0}: Error finding container 4d3fc474e2852c473505048a03f73d283a85e3e1e3f108988bdce1c714472460: Status 404 returned error can't find the container with id 4d3fc474e2852c473505048a03f73d283a85e3e1e3f108988bdce1c714472460 Oct 07 14:07:01 crc kubenswrapper[4959]: I1007 14:07:01.518080 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"439982d5-b039-45c1-9300-f39ce049cf33","Type":"ContainerStarted","Data":"739c6f0195e90152edc70dea4fdcbb93cf416f0187ffa45baefdf63a6b14f2ac"} Oct 07 14:07:01 crc kubenswrapper[4959]: I1007 14:07:01.525064 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0265cd4e-529d-40a6-a14c-6f39d6068633","Type":"ContainerStarted","Data":"4d3fc474e2852c473505048a03f73d283a85e3e1e3f108988bdce1c714472460"} Oct 07 14:07:02 crc kubenswrapper[4959]: I1007 14:07:02.536288 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"439982d5-b039-45c1-9300-f39ce049cf33","Type":"ContainerStarted","Data":"c5d23e4880832e9638d839b4bdcaa8a0c7dfcf7cc6189d3a723071dc3b26d038"} Oct 07 14:07:03 crc kubenswrapper[4959]: I1007 14:07:03.296621 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64fb5d8fd7-2tsgc"] Oct 07 14:07:03 crc kubenswrapper[4959]: I1007 14:07:03.298426 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" Oct 07 14:07:03 crc kubenswrapper[4959]: I1007 14:07:03.300746 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 07 14:07:03 crc kubenswrapper[4959]: I1007 14:07:03.310161 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64fb5d8fd7-2tsgc"] Oct 07 14:07:03 crc kubenswrapper[4959]: I1007 14:07:03.385024 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-ovsdbserver-sb\") pod \"dnsmasq-dns-64fb5d8fd7-2tsgc\" (UID: \"9df8a90a-5a19-48cf-90a4-05db2e51062a\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" Oct 07 14:07:03 crc kubenswrapper[4959]: I1007 14:07:03.385088 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxltf\" (UniqueName: \"kubernetes.io/projected/9df8a90a-5a19-48cf-90a4-05db2e51062a-kube-api-access-bxltf\") pod \"dnsmasq-dns-64fb5d8fd7-2tsgc\" (UID: \"9df8a90a-5a19-48cf-90a4-05db2e51062a\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" Oct 07 14:07:03 crc kubenswrapper[4959]: I1007 14:07:03.385242 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-openstack-edpm-ipam\") pod \"dnsmasq-dns-64fb5d8fd7-2tsgc\" (UID: \"9df8a90a-5a19-48cf-90a4-05db2e51062a\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" Oct 07 14:07:03 crc kubenswrapper[4959]: I1007 14:07:03.385326 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-dns-svc\") pod \"dnsmasq-dns-64fb5d8fd7-2tsgc\" (UID: \"9df8a90a-5a19-48cf-90a4-05db2e51062a\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" Oct 07 14:07:03 crc kubenswrapper[4959]: I1007 14:07:03.385486 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-config\") pod \"dnsmasq-dns-64fb5d8fd7-2tsgc\" (UID: \"9df8a90a-5a19-48cf-90a4-05db2e51062a\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" Oct 07 14:07:03 crc kubenswrapper[4959]: I1007 14:07:03.385546 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-ovsdbserver-nb\") pod \"dnsmasq-dns-64fb5d8fd7-2tsgc\" (UID: \"9df8a90a-5a19-48cf-90a4-05db2e51062a\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" Oct 07 14:07:03 crc kubenswrapper[4959]: I1007 14:07:03.486862 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxltf\" (UniqueName: \"kubernetes.io/projected/9df8a90a-5a19-48cf-90a4-05db2e51062a-kube-api-access-bxltf\") pod \"dnsmasq-dns-64fb5d8fd7-2tsgc\" (UID: \"9df8a90a-5a19-48cf-90a4-05db2e51062a\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" Oct 07 14:07:03 crc kubenswrapper[4959]: I1007 14:07:03.486962 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-openstack-edpm-ipam\") pod \"dnsmasq-dns-64fb5d8fd7-2tsgc\" (UID: \"9df8a90a-5a19-48cf-90a4-05db2e51062a\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" Oct 07 14:07:03 crc kubenswrapper[4959]: I1007 14:07:03.486994 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-dns-svc\") pod \"dnsmasq-dns-64fb5d8fd7-2tsgc\" (UID: \"9df8a90a-5a19-48cf-90a4-05db2e51062a\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" Oct 07 14:07:03 crc kubenswrapper[4959]: I1007 14:07:03.487053 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-config\") pod \"dnsmasq-dns-64fb5d8fd7-2tsgc\" (UID: \"9df8a90a-5a19-48cf-90a4-05db2e51062a\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" Oct 07 14:07:03 crc kubenswrapper[4959]: I1007 14:07:03.487090 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-ovsdbserver-nb\") pod \"dnsmasq-dns-64fb5d8fd7-2tsgc\" (UID: \"9df8a90a-5a19-48cf-90a4-05db2e51062a\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" Oct 07 14:07:03 crc kubenswrapper[4959]: I1007 14:07:03.487159 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-ovsdbserver-sb\") pod \"dnsmasq-dns-64fb5d8fd7-2tsgc\" (UID: \"9df8a90a-5a19-48cf-90a4-05db2e51062a\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" Oct 07 14:07:03 crc kubenswrapper[4959]: I1007 14:07:03.487896 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-openstack-edpm-ipam\") pod \"dnsmasq-dns-64fb5d8fd7-2tsgc\" (UID: \"9df8a90a-5a19-48cf-90a4-05db2e51062a\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" Oct 07 14:07:03 crc kubenswrapper[4959]: I1007 14:07:03.488182 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-dns-svc\") pod \"dnsmasq-dns-64fb5d8fd7-2tsgc\" (UID: \"9df8a90a-5a19-48cf-90a4-05db2e51062a\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" Oct 07 14:07:03 crc kubenswrapper[4959]: I1007 14:07:03.488190 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-ovsdbserver-sb\") pod \"dnsmasq-dns-64fb5d8fd7-2tsgc\" (UID: \"9df8a90a-5a19-48cf-90a4-05db2e51062a\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" Oct 07 14:07:03 crc kubenswrapper[4959]: I1007 14:07:03.488554 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-ovsdbserver-nb\") pod \"dnsmasq-dns-64fb5d8fd7-2tsgc\" (UID: \"9df8a90a-5a19-48cf-90a4-05db2e51062a\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" Oct 07 14:07:03 crc kubenswrapper[4959]: I1007 14:07:03.488765 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-config\") pod \"dnsmasq-dns-64fb5d8fd7-2tsgc\" (UID: \"9df8a90a-5a19-48cf-90a4-05db2e51062a\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" Oct 07 14:07:03 crc kubenswrapper[4959]: I1007 14:07:03.508243 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxltf\" (UniqueName: \"kubernetes.io/projected/9df8a90a-5a19-48cf-90a4-05db2e51062a-kube-api-access-bxltf\") pod \"dnsmasq-dns-64fb5d8fd7-2tsgc\" (UID: \"9df8a90a-5a19-48cf-90a4-05db2e51062a\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" Oct 07 14:07:03 crc kubenswrapper[4959]: I1007 14:07:03.546760 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0265cd4e-529d-40a6-a14c-6f39d6068633","Type":"ContainerStarted","Data":"ad26facb435eeefd1964e2f1b04169cd37651fe81ab7a945fbb701f1e21e5ef1"} Oct 07 14:07:03 crc kubenswrapper[4959]: I1007 14:07:03.618212 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" Oct 07 14:07:03 crc kubenswrapper[4959]: I1007 14:07:03.664070 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="90af7ce4-adf2-4722-a726-cc4d7d29056f" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.102:5671: i/o timeout" Oct 07 14:07:04 crc kubenswrapper[4959]: I1007 14:07:04.057534 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64fb5d8fd7-2tsgc"] Oct 07 14:07:04 crc kubenswrapper[4959]: I1007 14:07:04.557931 4959 generic.go:334] "Generic (PLEG): container finished" podID="9df8a90a-5a19-48cf-90a4-05db2e51062a" containerID="41019f5de636b7ab8dc3f7b113c3f2a630180dea751a73105bcf02e726cd76e2" exitCode=0 Oct 07 14:07:04 crc kubenswrapper[4959]: I1007 14:07:04.558022 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" event={"ID":"9df8a90a-5a19-48cf-90a4-05db2e51062a","Type":"ContainerDied","Data":"41019f5de636b7ab8dc3f7b113c3f2a630180dea751a73105bcf02e726cd76e2"} Oct 07 14:07:04 crc kubenswrapper[4959]: I1007 14:07:04.558054 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" event={"ID":"9df8a90a-5a19-48cf-90a4-05db2e51062a","Type":"ContainerStarted","Data":"2310149673fc269e8cc63855d4ceb73016c1e838b640ef0289cd55d0cd97c1a6"} Oct 07 14:07:05 crc kubenswrapper[4959]: I1007 14:07:05.568435 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" event={"ID":"9df8a90a-5a19-48cf-90a4-05db2e51062a","Type":"ContainerStarted","Data":"e57d688a13e9e5378596c8ea09c6f64e4cfb263dbe8d07d68d573555eefd6155"} Oct 07 14:07:05 crc kubenswrapper[4959]: I1007 14:07:05.568890 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" Oct 07 14:07:05 crc kubenswrapper[4959]: I1007 14:07:05.597884 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" podStartSLOduration=2.597851148 podStartE2EDuration="2.597851148s" podCreationTimestamp="2025-10-07 14:07:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:07:05.593603622 +0000 UTC m=+1307.677007957" watchObservedRunningTime="2025-10-07 14:07:05.597851148 +0000 UTC m=+1307.681255463" Oct 07 14:07:13 crc kubenswrapper[4959]: I1007 14:07:13.620295 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" Oct 07 14:07:13 crc kubenswrapper[4959]: I1007 14:07:13.743371 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-665946c669-2wpls"] Oct 07 14:07:13 crc kubenswrapper[4959]: I1007 14:07:13.743917 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-665946c669-2wpls" podUID="a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a" containerName="dnsmasq-dns" containerID="cri-o://527e7a4af5611fbce86419259a551f2f2efc05cc81b0f1179c3e5c0a1d9dd965" gracePeriod=10 Oct 07 14:07:13 crc kubenswrapper[4959]: I1007 14:07:13.852496 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-dbd9bf859-chpc6"] Oct 07 14:07:13 crc kubenswrapper[4959]: I1007 14:07:13.854273 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" Oct 07 14:07:13 crc kubenswrapper[4959]: I1007 14:07:13.877952 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dbd9bf859-chpc6"] Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.002167 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-ovsdbserver-nb\") pod \"dnsmasq-dns-dbd9bf859-chpc6\" (UID: \"2726035d-3257-42b4-b38a-8384e1db7d95\") " pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.002213 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-dns-svc\") pod \"dnsmasq-dns-dbd9bf859-chpc6\" (UID: \"2726035d-3257-42b4-b38a-8384e1db7d95\") " pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.002252 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmpdr\" (UniqueName: \"kubernetes.io/projected/2726035d-3257-42b4-b38a-8384e1db7d95-kube-api-access-pmpdr\") pod \"dnsmasq-dns-dbd9bf859-chpc6\" (UID: \"2726035d-3257-42b4-b38a-8384e1db7d95\") " pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.002299 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-config\") pod \"dnsmasq-dns-dbd9bf859-chpc6\" (UID: \"2726035d-3257-42b4-b38a-8384e1db7d95\") " pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.002323 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-openstack-edpm-ipam\") pod \"dnsmasq-dns-dbd9bf859-chpc6\" (UID: \"2726035d-3257-42b4-b38a-8384e1db7d95\") " pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.002400 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-ovsdbserver-sb\") pod \"dnsmasq-dns-dbd9bf859-chpc6\" (UID: \"2726035d-3257-42b4-b38a-8384e1db7d95\") " pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.103666 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-openstack-edpm-ipam\") pod \"dnsmasq-dns-dbd9bf859-chpc6\" (UID: \"2726035d-3257-42b4-b38a-8384e1db7d95\") " pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.104012 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-ovsdbserver-sb\") pod \"dnsmasq-dns-dbd9bf859-chpc6\" (UID: \"2726035d-3257-42b4-b38a-8384e1db7d95\") " pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.104130 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-ovsdbserver-nb\") pod \"dnsmasq-dns-dbd9bf859-chpc6\" (UID: \"2726035d-3257-42b4-b38a-8384e1db7d95\") " pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.104158 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-dns-svc\") pod \"dnsmasq-dns-dbd9bf859-chpc6\" (UID: \"2726035d-3257-42b4-b38a-8384e1db7d95\") " pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.104195 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmpdr\" (UniqueName: \"kubernetes.io/projected/2726035d-3257-42b4-b38a-8384e1db7d95-kube-api-access-pmpdr\") pod \"dnsmasq-dns-dbd9bf859-chpc6\" (UID: \"2726035d-3257-42b4-b38a-8384e1db7d95\") " pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.104224 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-config\") pod \"dnsmasq-dns-dbd9bf859-chpc6\" (UID: \"2726035d-3257-42b4-b38a-8384e1db7d95\") " pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.104810 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-dns-svc\") pod \"dnsmasq-dns-dbd9bf859-chpc6\" (UID: \"2726035d-3257-42b4-b38a-8384e1db7d95\") " pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.104878 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-openstack-edpm-ipam\") pod \"dnsmasq-dns-dbd9bf859-chpc6\" (UID: \"2726035d-3257-42b4-b38a-8384e1db7d95\") " pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.104967 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-ovsdbserver-nb\") pod \"dnsmasq-dns-dbd9bf859-chpc6\" (UID: \"2726035d-3257-42b4-b38a-8384e1db7d95\") " pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.105045 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-config\") pod \"dnsmasq-dns-dbd9bf859-chpc6\" (UID: \"2726035d-3257-42b4-b38a-8384e1db7d95\") " pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.105618 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-ovsdbserver-sb\") pod \"dnsmasq-dns-dbd9bf859-chpc6\" (UID: \"2726035d-3257-42b4-b38a-8384e1db7d95\") " pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.124123 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmpdr\" (UniqueName: \"kubernetes.io/projected/2726035d-3257-42b4-b38a-8384e1db7d95-kube-api-access-pmpdr\") pod \"dnsmasq-dns-dbd9bf859-chpc6\" (UID: \"2726035d-3257-42b4-b38a-8384e1db7d95\") " pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.215529 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-665946c669-2wpls" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.221826 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.307190 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-dns-svc\") pod \"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a\" (UID: \"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a\") " Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.307416 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-ovsdbserver-nb\") pod \"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a\" (UID: \"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a\") " Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.307470 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nd6l\" (UniqueName: \"kubernetes.io/projected/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-kube-api-access-6nd6l\") pod \"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a\" (UID: \"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a\") " Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.307553 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-config\") pod \"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a\" (UID: \"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a\") " Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.307595 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-ovsdbserver-sb\") pod \"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a\" (UID: \"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a\") " Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.317865 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-kube-api-access-6nd6l" (OuterVolumeSpecName: "kube-api-access-6nd6l") pod "a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a" (UID: "a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a"). InnerVolumeSpecName "kube-api-access-6nd6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.384693 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a" (UID: "a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.409222 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nd6l\" (UniqueName: \"kubernetes.io/projected/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-kube-api-access-6nd6l\") on node \"crc\" DevicePath \"\"" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.409257 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.411829 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a" (UID: "a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.416815 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a" (UID: "a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.449337 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-config" (OuterVolumeSpecName: "config") pod "a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a" (UID: "a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.516094 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-config\") on node \"crc\" DevicePath \"\"" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.516145 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.516157 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.661167 4959 generic.go:334] "Generic (PLEG): container finished" podID="a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a" containerID="527e7a4af5611fbce86419259a551f2f2efc05cc81b0f1179c3e5c0a1d9dd965" exitCode=0 Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.661273 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-665946c669-2wpls" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.677527 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-665946c669-2wpls" event={"ID":"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a","Type":"ContainerDied","Data":"527e7a4af5611fbce86419259a551f2f2efc05cc81b0f1179c3e5c0a1d9dd965"} Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.677580 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-665946c669-2wpls" event={"ID":"a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a","Type":"ContainerDied","Data":"8cef8f2bd89b8544ebbd654c9e764fc41b79ec608e297d827ba08744df44c220"} Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.677603 4959 scope.go:117] "RemoveContainer" containerID="527e7a4af5611fbce86419259a551f2f2efc05cc81b0f1179c3e5c0a1d9dd965" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.708078 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-665946c669-2wpls"] Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.714986 4959 scope.go:117] "RemoveContainer" containerID="effeeb87aac91c6f207ee097e5ddf40065b20990f8649ac951a5bdbd6b4ba804" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.716760 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-665946c669-2wpls"] Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.726867 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dbd9bf859-chpc6"] Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.735343 4959 scope.go:117] "RemoveContainer" containerID="527e7a4af5611fbce86419259a551f2f2efc05cc81b0f1179c3e5c0a1d9dd965" Oct 07 14:07:14 crc kubenswrapper[4959]: E1007 14:07:14.735763 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"527e7a4af5611fbce86419259a551f2f2efc05cc81b0f1179c3e5c0a1d9dd965\": container with ID starting with 527e7a4af5611fbce86419259a551f2f2efc05cc81b0f1179c3e5c0a1d9dd965 not found: ID does not exist" containerID="527e7a4af5611fbce86419259a551f2f2efc05cc81b0f1179c3e5c0a1d9dd965" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.735826 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"527e7a4af5611fbce86419259a551f2f2efc05cc81b0f1179c3e5c0a1d9dd965"} err="failed to get container status \"527e7a4af5611fbce86419259a551f2f2efc05cc81b0f1179c3e5c0a1d9dd965\": rpc error: code = NotFound desc = could not find container \"527e7a4af5611fbce86419259a551f2f2efc05cc81b0f1179c3e5c0a1d9dd965\": container with ID starting with 527e7a4af5611fbce86419259a551f2f2efc05cc81b0f1179c3e5c0a1d9dd965 not found: ID does not exist" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.735860 4959 scope.go:117] "RemoveContainer" containerID="effeeb87aac91c6f207ee097e5ddf40065b20990f8649ac951a5bdbd6b4ba804" Oct 07 14:07:14 crc kubenswrapper[4959]: E1007 14:07:14.736135 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"effeeb87aac91c6f207ee097e5ddf40065b20990f8649ac951a5bdbd6b4ba804\": container with ID starting with effeeb87aac91c6f207ee097e5ddf40065b20990f8649ac951a5bdbd6b4ba804 not found: ID does not exist" containerID="effeeb87aac91c6f207ee097e5ddf40065b20990f8649ac951a5bdbd6b4ba804" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.736157 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"effeeb87aac91c6f207ee097e5ddf40065b20990f8649ac951a5bdbd6b4ba804"} err="failed to get container status \"effeeb87aac91c6f207ee097e5ddf40065b20990f8649ac951a5bdbd6b4ba804\": rpc error: code = NotFound desc = could not find container \"effeeb87aac91c6f207ee097e5ddf40065b20990f8649ac951a5bdbd6b4ba804\": container with ID starting with effeeb87aac91c6f207ee097e5ddf40065b20990f8649ac951a5bdbd6b4ba804 not found: ID does not exist" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.947653 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf"] Oct 07 14:07:14 crc kubenswrapper[4959]: E1007 14:07:14.948298 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a" containerName="init" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.948314 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a" containerName="init" Oct 07 14:07:14 crc kubenswrapper[4959]: E1007 14:07:14.948338 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a" containerName="dnsmasq-dns" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.948344 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a" containerName="dnsmasq-dns" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.948526 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a" containerName="dnsmasq-dns" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.949118 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.950898 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-c6vmh" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.951657 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.951781 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.952247 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 14:07:14 crc kubenswrapper[4959]: I1007 14:07:14.959508 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf"] Oct 07 14:07:15 crc kubenswrapper[4959]: I1007 14:07:15.125957 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc5887d2-a870-4086-b94b-4abdfefd42bd-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf\" (UID: \"bc5887d2-a870-4086-b94b-4abdfefd42bd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf" Oct 07 14:07:15 crc kubenswrapper[4959]: I1007 14:07:15.126536 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc5887d2-a870-4086-b94b-4abdfefd42bd-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf\" (UID: \"bc5887d2-a870-4086-b94b-4abdfefd42bd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf" Oct 07 14:07:15 crc kubenswrapper[4959]: I1007 14:07:15.126625 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc5887d2-a870-4086-b94b-4abdfefd42bd-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf\" (UID: \"bc5887d2-a870-4086-b94b-4abdfefd42bd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf" Oct 07 14:07:15 crc kubenswrapper[4959]: I1007 14:07:15.126843 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54q7k\" (UniqueName: \"kubernetes.io/projected/bc5887d2-a870-4086-b94b-4abdfefd42bd-kube-api-access-54q7k\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf\" (UID: \"bc5887d2-a870-4086-b94b-4abdfefd42bd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf" Oct 07 14:07:15 crc kubenswrapper[4959]: I1007 14:07:15.228603 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc5887d2-a870-4086-b94b-4abdfefd42bd-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf\" (UID: \"bc5887d2-a870-4086-b94b-4abdfefd42bd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf" Oct 07 14:07:15 crc kubenswrapper[4959]: I1007 14:07:15.228703 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc5887d2-a870-4086-b94b-4abdfefd42bd-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf\" (UID: \"bc5887d2-a870-4086-b94b-4abdfefd42bd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf" Oct 07 14:07:15 crc kubenswrapper[4959]: I1007 14:07:15.228832 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54q7k\" (UniqueName: \"kubernetes.io/projected/bc5887d2-a870-4086-b94b-4abdfefd42bd-kube-api-access-54q7k\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf\" (UID: \"bc5887d2-a870-4086-b94b-4abdfefd42bd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf" Oct 07 14:07:15 crc kubenswrapper[4959]: I1007 14:07:15.228952 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc5887d2-a870-4086-b94b-4abdfefd42bd-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf\" (UID: \"bc5887d2-a870-4086-b94b-4abdfefd42bd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf" Oct 07 14:07:15 crc kubenswrapper[4959]: I1007 14:07:15.233411 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc5887d2-a870-4086-b94b-4abdfefd42bd-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf\" (UID: \"bc5887d2-a870-4086-b94b-4abdfefd42bd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf" Oct 07 14:07:15 crc kubenswrapper[4959]: I1007 14:07:15.233746 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc5887d2-a870-4086-b94b-4abdfefd42bd-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf\" (UID: \"bc5887d2-a870-4086-b94b-4abdfefd42bd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf" Oct 07 14:07:15 crc kubenswrapper[4959]: I1007 14:07:15.233799 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc5887d2-a870-4086-b94b-4abdfefd42bd-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf\" (UID: \"bc5887d2-a870-4086-b94b-4abdfefd42bd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf" Oct 07 14:07:15 crc kubenswrapper[4959]: I1007 14:07:15.244522 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54q7k\" (UniqueName: \"kubernetes.io/projected/bc5887d2-a870-4086-b94b-4abdfefd42bd-kube-api-access-54q7k\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf\" (UID: \"bc5887d2-a870-4086-b94b-4abdfefd42bd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf" Oct 07 14:07:15 crc kubenswrapper[4959]: I1007 14:07:15.333694 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf" Oct 07 14:07:15 crc kubenswrapper[4959]: I1007 14:07:15.672412 4959 generic.go:334] "Generic (PLEG): container finished" podID="2726035d-3257-42b4-b38a-8384e1db7d95" containerID="c3f231de1ecd6f82555d9b5f33486a10f5348c3e86494bed7ff2d82de652ca65" exitCode=0 Oct 07 14:07:15 crc kubenswrapper[4959]: I1007 14:07:15.672518 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" event={"ID":"2726035d-3257-42b4-b38a-8384e1db7d95","Type":"ContainerDied","Data":"c3f231de1ecd6f82555d9b5f33486a10f5348c3e86494bed7ff2d82de652ca65"} Oct 07 14:07:15 crc kubenswrapper[4959]: I1007 14:07:15.672816 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" event={"ID":"2726035d-3257-42b4-b38a-8384e1db7d95","Type":"ContainerStarted","Data":"b98de7af5210491eaf1282c45f8431e456445aa6fafd4f71c1f30a0da388eafd"} Oct 07 14:07:15 crc kubenswrapper[4959]: I1007 14:07:15.928197 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf"] Oct 07 14:07:15 crc kubenswrapper[4959]: W1007 14:07:15.932783 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc5887d2_a870_4086_b94b_4abdfefd42bd.slice/crio-9f0639ea37d979474add178ddce22c50f16c63c1dd77c29d929b7a5997b81293 WatchSource:0}: Error finding container 9f0639ea37d979474add178ddce22c50f16c63c1dd77c29d929b7a5997b81293: Status 404 returned error can't find the container with id 9f0639ea37d979474add178ddce22c50f16c63c1dd77c29d929b7a5997b81293 Oct 07 14:07:16 crc kubenswrapper[4959]: I1007 14:07:16.687867 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a" path="/var/lib/kubelet/pods/a2e9e1ab-2ed3-4fa6-8d20-669cc7088b1a/volumes" Oct 07 14:07:16 crc kubenswrapper[4959]: I1007 14:07:16.697631 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" event={"ID":"2726035d-3257-42b4-b38a-8384e1db7d95","Type":"ContainerStarted","Data":"7c8d9601815a9dd378f0655b175e064b752834b59a1de3d9f0474dbdf08e7dad"} Oct 07 14:07:16 crc kubenswrapper[4959]: I1007 14:07:16.697787 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" Oct 07 14:07:16 crc kubenswrapper[4959]: I1007 14:07:16.699478 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf" event={"ID":"bc5887d2-a870-4086-b94b-4abdfefd42bd","Type":"ContainerStarted","Data":"9f0639ea37d979474add178ddce22c50f16c63c1dd77c29d929b7a5997b81293"} Oct 07 14:07:16 crc kubenswrapper[4959]: I1007 14:07:16.721404 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" podStartSLOduration=3.721387111 podStartE2EDuration="3.721387111s" podCreationTimestamp="2025-10-07 14:07:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:07:16.719554385 +0000 UTC m=+1318.802958720" watchObservedRunningTime="2025-10-07 14:07:16.721387111 +0000 UTC m=+1318.804791436" Oct 07 14:07:20 crc kubenswrapper[4959]: I1007 14:07:20.630407 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:07:20 crc kubenswrapper[4959]: I1007 14:07:20.631643 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:07:23 crc kubenswrapper[4959]: I1007 14:07:23.946069 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 14:07:24 crc kubenswrapper[4959]: I1007 14:07:24.223200 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" Oct 07 14:07:24 crc kubenswrapper[4959]: I1007 14:07:24.322491 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64fb5d8fd7-2tsgc"] Oct 07 14:07:24 crc kubenswrapper[4959]: I1007 14:07:24.322712 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" podUID="9df8a90a-5a19-48cf-90a4-05db2e51062a" containerName="dnsmasq-dns" containerID="cri-o://e57d688a13e9e5378596c8ea09c6f64e4cfb263dbe8d07d68d573555eefd6155" gracePeriod=10 Oct 07 14:07:24 crc kubenswrapper[4959]: I1007 14:07:24.741475 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" Oct 07 14:07:24 crc kubenswrapper[4959]: I1007 14:07:24.773894 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf" event={"ID":"bc5887d2-a870-4086-b94b-4abdfefd42bd","Type":"ContainerStarted","Data":"eb5fe6d5b2fa51a8cbb00614d1816c41eb1ca032e53c3477ddfa2f2a340b2a0a"} Oct 07 14:07:24 crc kubenswrapper[4959]: I1007 14:07:24.776692 4959 generic.go:334] "Generic (PLEG): container finished" podID="9df8a90a-5a19-48cf-90a4-05db2e51062a" containerID="e57d688a13e9e5378596c8ea09c6f64e4cfb263dbe8d07d68d573555eefd6155" exitCode=0 Oct 07 14:07:24 crc kubenswrapper[4959]: I1007 14:07:24.776750 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" Oct 07 14:07:24 crc kubenswrapper[4959]: I1007 14:07:24.776748 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" event={"ID":"9df8a90a-5a19-48cf-90a4-05db2e51062a","Type":"ContainerDied","Data":"e57d688a13e9e5378596c8ea09c6f64e4cfb263dbe8d07d68d573555eefd6155"} Oct 07 14:07:24 crc kubenswrapper[4959]: I1007 14:07:24.776917 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64fb5d8fd7-2tsgc" event={"ID":"9df8a90a-5a19-48cf-90a4-05db2e51062a","Type":"ContainerDied","Data":"2310149673fc269e8cc63855d4ceb73016c1e838b640ef0289cd55d0cd97c1a6"} Oct 07 14:07:24 crc kubenswrapper[4959]: I1007 14:07:24.776934 4959 scope.go:117] "RemoveContainer" containerID="e57d688a13e9e5378596c8ea09c6f64e4cfb263dbe8d07d68d573555eefd6155" Oct 07 14:07:24 crc kubenswrapper[4959]: I1007 14:07:24.802510 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf" podStartSLOduration=2.7943529700000003 podStartE2EDuration="10.802485104s" podCreationTimestamp="2025-10-07 14:07:14 +0000 UTC" firstStartedPulling="2025-10-07 14:07:15.935705812 +0000 UTC m=+1318.019110137" lastFinishedPulling="2025-10-07 14:07:23.943837946 +0000 UTC m=+1326.027242271" observedRunningTime="2025-10-07 14:07:24.792489544 +0000 UTC m=+1326.875893869" watchObservedRunningTime="2025-10-07 14:07:24.802485104 +0000 UTC m=+1326.885889429" Oct 07 14:07:24 crc kubenswrapper[4959]: I1007 14:07:24.805293 4959 scope.go:117] "RemoveContainer" containerID="41019f5de636b7ab8dc3f7b113c3f2a630180dea751a73105bcf02e726cd76e2" Oct 07 14:07:24 crc kubenswrapper[4959]: I1007 14:07:24.826848 4959 scope.go:117] "RemoveContainer" containerID="e57d688a13e9e5378596c8ea09c6f64e4cfb263dbe8d07d68d573555eefd6155" Oct 07 14:07:24 crc kubenswrapper[4959]: E1007 14:07:24.827276 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e57d688a13e9e5378596c8ea09c6f64e4cfb263dbe8d07d68d573555eefd6155\": container with ID starting with e57d688a13e9e5378596c8ea09c6f64e4cfb263dbe8d07d68d573555eefd6155 not found: ID does not exist" containerID="e57d688a13e9e5378596c8ea09c6f64e4cfb263dbe8d07d68d573555eefd6155" Oct 07 14:07:24 crc kubenswrapper[4959]: I1007 14:07:24.827308 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e57d688a13e9e5378596c8ea09c6f64e4cfb263dbe8d07d68d573555eefd6155"} err="failed to get container status \"e57d688a13e9e5378596c8ea09c6f64e4cfb263dbe8d07d68d573555eefd6155\": rpc error: code = NotFound desc = could not find container \"e57d688a13e9e5378596c8ea09c6f64e4cfb263dbe8d07d68d573555eefd6155\": container with ID starting with e57d688a13e9e5378596c8ea09c6f64e4cfb263dbe8d07d68d573555eefd6155 not found: ID does not exist" Oct 07 14:07:24 crc kubenswrapper[4959]: I1007 14:07:24.827332 4959 scope.go:117] "RemoveContainer" containerID="41019f5de636b7ab8dc3f7b113c3f2a630180dea751a73105bcf02e726cd76e2" Oct 07 14:07:24 crc kubenswrapper[4959]: E1007 14:07:24.827537 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41019f5de636b7ab8dc3f7b113c3f2a630180dea751a73105bcf02e726cd76e2\": container with ID starting with 41019f5de636b7ab8dc3f7b113c3f2a630180dea751a73105bcf02e726cd76e2 not found: ID does not exist" containerID="41019f5de636b7ab8dc3f7b113c3f2a630180dea751a73105bcf02e726cd76e2" Oct 07 14:07:24 crc kubenswrapper[4959]: I1007 14:07:24.827557 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41019f5de636b7ab8dc3f7b113c3f2a630180dea751a73105bcf02e726cd76e2"} err="failed to get container status \"41019f5de636b7ab8dc3f7b113c3f2a630180dea751a73105bcf02e726cd76e2\": rpc error: code = NotFound desc = could not find container \"41019f5de636b7ab8dc3f7b113c3f2a630180dea751a73105bcf02e726cd76e2\": container with ID starting with 41019f5de636b7ab8dc3f7b113c3f2a630180dea751a73105bcf02e726cd76e2 not found: ID does not exist" Oct 07 14:07:24 crc kubenswrapper[4959]: I1007 14:07:24.918696 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-dns-svc\") pod \"9df8a90a-5a19-48cf-90a4-05db2e51062a\" (UID: \"9df8a90a-5a19-48cf-90a4-05db2e51062a\") " Oct 07 14:07:24 crc kubenswrapper[4959]: I1007 14:07:24.918883 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-ovsdbserver-nb\") pod \"9df8a90a-5a19-48cf-90a4-05db2e51062a\" (UID: \"9df8a90a-5a19-48cf-90a4-05db2e51062a\") " Oct 07 14:07:24 crc kubenswrapper[4959]: I1007 14:07:24.919741 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-ovsdbserver-sb\") pod \"9df8a90a-5a19-48cf-90a4-05db2e51062a\" (UID: \"9df8a90a-5a19-48cf-90a4-05db2e51062a\") " Oct 07 14:07:24 crc kubenswrapper[4959]: I1007 14:07:24.919817 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxltf\" (UniqueName: \"kubernetes.io/projected/9df8a90a-5a19-48cf-90a4-05db2e51062a-kube-api-access-bxltf\") pod \"9df8a90a-5a19-48cf-90a4-05db2e51062a\" (UID: \"9df8a90a-5a19-48cf-90a4-05db2e51062a\") " Oct 07 14:07:24 crc kubenswrapper[4959]: I1007 14:07:24.919866 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-config\") pod \"9df8a90a-5a19-48cf-90a4-05db2e51062a\" (UID: \"9df8a90a-5a19-48cf-90a4-05db2e51062a\") " Oct 07 14:07:24 crc kubenswrapper[4959]: I1007 14:07:24.919957 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-openstack-edpm-ipam\") pod \"9df8a90a-5a19-48cf-90a4-05db2e51062a\" (UID: \"9df8a90a-5a19-48cf-90a4-05db2e51062a\") " Oct 07 14:07:24 crc kubenswrapper[4959]: I1007 14:07:24.924633 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9df8a90a-5a19-48cf-90a4-05db2e51062a-kube-api-access-bxltf" (OuterVolumeSpecName: "kube-api-access-bxltf") pod "9df8a90a-5a19-48cf-90a4-05db2e51062a" (UID: "9df8a90a-5a19-48cf-90a4-05db2e51062a"). InnerVolumeSpecName "kube-api-access-bxltf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:07:24 crc kubenswrapper[4959]: I1007 14:07:24.966416 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "9df8a90a-5a19-48cf-90a4-05db2e51062a" (UID: "9df8a90a-5a19-48cf-90a4-05db2e51062a"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:07:24 crc kubenswrapper[4959]: I1007 14:07:24.969839 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9df8a90a-5a19-48cf-90a4-05db2e51062a" (UID: "9df8a90a-5a19-48cf-90a4-05db2e51062a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:07:24 crc kubenswrapper[4959]: I1007 14:07:24.970545 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-config" (OuterVolumeSpecName: "config") pod "9df8a90a-5a19-48cf-90a4-05db2e51062a" (UID: "9df8a90a-5a19-48cf-90a4-05db2e51062a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:07:24 crc kubenswrapper[4959]: I1007 14:07:24.973302 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9df8a90a-5a19-48cf-90a4-05db2e51062a" (UID: "9df8a90a-5a19-48cf-90a4-05db2e51062a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:07:24 crc kubenswrapper[4959]: I1007 14:07:24.974768 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9df8a90a-5a19-48cf-90a4-05db2e51062a" (UID: "9df8a90a-5a19-48cf-90a4-05db2e51062a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:07:25 crc kubenswrapper[4959]: I1007 14:07:25.022057 4959 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 07 14:07:25 crc kubenswrapper[4959]: I1007 14:07:25.022090 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 14:07:25 crc kubenswrapper[4959]: I1007 14:07:25.022110 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 14:07:25 crc kubenswrapper[4959]: I1007 14:07:25.022119 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 14:07:25 crc kubenswrapper[4959]: I1007 14:07:25.022128 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxltf\" (UniqueName: \"kubernetes.io/projected/9df8a90a-5a19-48cf-90a4-05db2e51062a-kube-api-access-bxltf\") on node \"crc\" DevicePath \"\"" Oct 07 14:07:25 crc kubenswrapper[4959]: I1007 14:07:25.022139 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9df8a90a-5a19-48cf-90a4-05db2e51062a-config\") on node \"crc\" DevicePath \"\"" Oct 07 14:07:25 crc kubenswrapper[4959]: I1007 14:07:25.120780 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64fb5d8fd7-2tsgc"] Oct 07 14:07:25 crc kubenswrapper[4959]: I1007 14:07:25.128484 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64fb5d8fd7-2tsgc"] Oct 07 14:07:26 crc kubenswrapper[4959]: I1007 14:07:26.663345 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9df8a90a-5a19-48cf-90a4-05db2e51062a" path="/var/lib/kubelet/pods/9df8a90a-5a19-48cf-90a4-05db2e51062a/volumes" Oct 07 14:07:34 crc kubenswrapper[4959]: I1007 14:07:34.885816 4959 generic.go:334] "Generic (PLEG): container finished" podID="439982d5-b039-45c1-9300-f39ce049cf33" containerID="c5d23e4880832e9638d839b4bdcaa8a0c7dfcf7cc6189d3a723071dc3b26d038" exitCode=0 Oct 07 14:07:34 crc kubenswrapper[4959]: I1007 14:07:34.885889 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"439982d5-b039-45c1-9300-f39ce049cf33","Type":"ContainerDied","Data":"c5d23e4880832e9638d839b4bdcaa8a0c7dfcf7cc6189d3a723071dc3b26d038"} Oct 07 14:07:35 crc kubenswrapper[4959]: I1007 14:07:35.898519 4959 generic.go:334] "Generic (PLEG): container finished" podID="0265cd4e-529d-40a6-a14c-6f39d6068633" containerID="ad26facb435eeefd1964e2f1b04169cd37651fe81ab7a945fbb701f1e21e5ef1" exitCode=0 Oct 07 14:07:35 crc kubenswrapper[4959]: I1007 14:07:35.898610 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0265cd4e-529d-40a6-a14c-6f39d6068633","Type":"ContainerDied","Data":"ad26facb435eeefd1964e2f1b04169cd37651fe81ab7a945fbb701f1e21e5ef1"} Oct 07 14:07:35 crc kubenswrapper[4959]: I1007 14:07:35.902581 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"439982d5-b039-45c1-9300-f39ce049cf33","Type":"ContainerStarted","Data":"ae20ea6b728e50a20d872412f591f8e66af163d7d6f684fcbbe1401b9fe10567"} Oct 07 14:07:35 crc kubenswrapper[4959]: I1007 14:07:35.902920 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 07 14:07:35 crc kubenswrapper[4959]: I1007 14:07:35.965633 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.965615342 podStartE2EDuration="36.965615342s" podCreationTimestamp="2025-10-07 14:06:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:07:35.962289109 +0000 UTC m=+1338.045693434" watchObservedRunningTime="2025-10-07 14:07:35.965615342 +0000 UTC m=+1338.049019667" Oct 07 14:07:36 crc kubenswrapper[4959]: I1007 14:07:36.911843 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0265cd4e-529d-40a6-a14c-6f39d6068633","Type":"ContainerStarted","Data":"d9095b5ba85da9447d87e4b4045568fa67557ca9347e921384c54eaff9de7e38"} Oct 07 14:07:36 crc kubenswrapper[4959]: I1007 14:07:36.912491 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:36 crc kubenswrapper[4959]: I1007 14:07:36.933883 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.933861728 podStartE2EDuration="36.933861728s" podCreationTimestamp="2025-10-07 14:07:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:07:36.930432302 +0000 UTC m=+1339.013836627" watchObservedRunningTime="2025-10-07 14:07:36.933861728 +0000 UTC m=+1339.017266073" Oct 07 14:07:37 crc kubenswrapper[4959]: I1007 14:07:37.920475 4959 generic.go:334] "Generic (PLEG): container finished" podID="bc5887d2-a870-4086-b94b-4abdfefd42bd" containerID="eb5fe6d5b2fa51a8cbb00614d1816c41eb1ca032e53c3477ddfa2f2a340b2a0a" exitCode=0 Oct 07 14:07:37 crc kubenswrapper[4959]: I1007 14:07:37.920693 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf" event={"ID":"bc5887d2-a870-4086-b94b-4abdfefd42bd","Type":"ContainerDied","Data":"eb5fe6d5b2fa51a8cbb00614d1816c41eb1ca032e53c3477ddfa2f2a340b2a0a"} Oct 07 14:07:39 crc kubenswrapper[4959]: I1007 14:07:39.301167 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf" Oct 07 14:07:39 crc kubenswrapper[4959]: I1007 14:07:39.395571 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54q7k\" (UniqueName: \"kubernetes.io/projected/bc5887d2-a870-4086-b94b-4abdfefd42bd-kube-api-access-54q7k\") pod \"bc5887d2-a870-4086-b94b-4abdfefd42bd\" (UID: \"bc5887d2-a870-4086-b94b-4abdfefd42bd\") " Oct 07 14:07:39 crc kubenswrapper[4959]: I1007 14:07:39.395773 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc5887d2-a870-4086-b94b-4abdfefd42bd-repo-setup-combined-ca-bundle\") pod \"bc5887d2-a870-4086-b94b-4abdfefd42bd\" (UID: \"bc5887d2-a870-4086-b94b-4abdfefd42bd\") " Oct 07 14:07:39 crc kubenswrapper[4959]: I1007 14:07:39.395811 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc5887d2-a870-4086-b94b-4abdfefd42bd-ssh-key\") pod \"bc5887d2-a870-4086-b94b-4abdfefd42bd\" (UID: \"bc5887d2-a870-4086-b94b-4abdfefd42bd\") " Oct 07 14:07:39 crc kubenswrapper[4959]: I1007 14:07:39.395893 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc5887d2-a870-4086-b94b-4abdfefd42bd-inventory\") pod \"bc5887d2-a870-4086-b94b-4abdfefd42bd\" (UID: \"bc5887d2-a870-4086-b94b-4abdfefd42bd\") " Oct 07 14:07:39 crc kubenswrapper[4959]: I1007 14:07:39.401555 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5887d2-a870-4086-b94b-4abdfefd42bd-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "bc5887d2-a870-4086-b94b-4abdfefd42bd" (UID: "bc5887d2-a870-4086-b94b-4abdfefd42bd"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:07:39 crc kubenswrapper[4959]: I1007 14:07:39.421652 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5887d2-a870-4086-b94b-4abdfefd42bd-kube-api-access-54q7k" (OuterVolumeSpecName: "kube-api-access-54q7k") pod "bc5887d2-a870-4086-b94b-4abdfefd42bd" (UID: "bc5887d2-a870-4086-b94b-4abdfefd42bd"). InnerVolumeSpecName "kube-api-access-54q7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:07:39 crc kubenswrapper[4959]: I1007 14:07:39.425354 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5887d2-a870-4086-b94b-4abdfefd42bd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bc5887d2-a870-4086-b94b-4abdfefd42bd" (UID: "bc5887d2-a870-4086-b94b-4abdfefd42bd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:07:39 crc kubenswrapper[4959]: I1007 14:07:39.436853 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5887d2-a870-4086-b94b-4abdfefd42bd-inventory" (OuterVolumeSpecName: "inventory") pod "bc5887d2-a870-4086-b94b-4abdfefd42bd" (UID: "bc5887d2-a870-4086-b94b-4abdfefd42bd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:07:39 crc kubenswrapper[4959]: I1007 14:07:39.498319 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54q7k\" (UniqueName: \"kubernetes.io/projected/bc5887d2-a870-4086-b94b-4abdfefd42bd-kube-api-access-54q7k\") on node \"crc\" DevicePath \"\"" Oct 07 14:07:39 crc kubenswrapper[4959]: I1007 14:07:39.498366 4959 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc5887d2-a870-4086-b94b-4abdfefd42bd-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:07:39 crc kubenswrapper[4959]: I1007 14:07:39.498377 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc5887d2-a870-4086-b94b-4abdfefd42bd-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 14:07:39 crc kubenswrapper[4959]: I1007 14:07:39.498387 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc5887d2-a870-4086-b94b-4abdfefd42bd-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 14:07:39 crc kubenswrapper[4959]: I1007 14:07:39.937430 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf" event={"ID":"bc5887d2-a870-4086-b94b-4abdfefd42bd","Type":"ContainerDied","Data":"9f0639ea37d979474add178ddce22c50f16c63c1dd77c29d929b7a5997b81293"} Oct 07 14:07:39 crc kubenswrapper[4959]: I1007 14:07:39.937772 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f0639ea37d979474add178ddce22c50f16c63c1dd77c29d929b7a5997b81293" Oct 07 14:07:39 crc kubenswrapper[4959]: I1007 14:07:39.937708 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf" Oct 07 14:07:40 crc kubenswrapper[4959]: I1007 14:07:40.015914 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4"] Oct 07 14:07:40 crc kubenswrapper[4959]: E1007 14:07:40.016680 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc5887d2-a870-4086-b94b-4abdfefd42bd" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 07 14:07:40 crc kubenswrapper[4959]: I1007 14:07:40.016809 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc5887d2-a870-4086-b94b-4abdfefd42bd" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 07 14:07:40 crc kubenswrapper[4959]: E1007 14:07:40.016941 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9df8a90a-5a19-48cf-90a4-05db2e51062a" containerName="dnsmasq-dns" Oct 07 14:07:40 crc kubenswrapper[4959]: I1007 14:07:40.017045 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="9df8a90a-5a19-48cf-90a4-05db2e51062a" containerName="dnsmasq-dns" Oct 07 14:07:40 crc kubenswrapper[4959]: E1007 14:07:40.017143 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9df8a90a-5a19-48cf-90a4-05db2e51062a" containerName="init" Oct 07 14:07:40 crc kubenswrapper[4959]: I1007 14:07:40.017234 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="9df8a90a-5a19-48cf-90a4-05db2e51062a" containerName="init" Oct 07 14:07:40 crc kubenswrapper[4959]: I1007 14:07:40.017475 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="9df8a90a-5a19-48cf-90a4-05db2e51062a" containerName="dnsmasq-dns" Oct 07 14:07:40 crc kubenswrapper[4959]: I1007 14:07:40.017546 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc5887d2-a870-4086-b94b-4abdfefd42bd" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 07 14:07:40 crc kubenswrapper[4959]: I1007 14:07:40.018280 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4" Oct 07 14:07:40 crc kubenswrapper[4959]: I1007 14:07:40.020027 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-c6vmh" Oct 07 14:07:40 crc kubenswrapper[4959]: I1007 14:07:40.020639 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 14:07:40 crc kubenswrapper[4959]: I1007 14:07:40.020697 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 14:07:40 crc kubenswrapper[4959]: I1007 14:07:40.025663 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4"] Oct 07 14:07:40 crc kubenswrapper[4959]: I1007 14:07:40.026322 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 14:07:40 crc kubenswrapper[4959]: I1007 14:07:40.109937 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxwcb\" (UniqueName: \"kubernetes.io/projected/e7967b3f-3a72-47d4-935d-4fa777b0fc46-kube-api-access-mxwcb\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4\" (UID: \"e7967b3f-3a72-47d4-935d-4fa777b0fc46\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4" Oct 07 14:07:40 crc kubenswrapper[4959]: I1007 14:07:40.110191 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e7967b3f-3a72-47d4-935d-4fa777b0fc46-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4\" (UID: \"e7967b3f-3a72-47d4-935d-4fa777b0fc46\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4" Oct 07 14:07:40 crc kubenswrapper[4959]: I1007 14:07:40.110267 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7967b3f-3a72-47d4-935d-4fa777b0fc46-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4\" (UID: \"e7967b3f-3a72-47d4-935d-4fa777b0fc46\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4" Oct 07 14:07:40 crc kubenswrapper[4959]: I1007 14:07:40.110374 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e7967b3f-3a72-47d4-935d-4fa777b0fc46-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4\" (UID: \"e7967b3f-3a72-47d4-935d-4fa777b0fc46\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4" Oct 07 14:07:40 crc kubenswrapper[4959]: I1007 14:07:40.212166 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7967b3f-3a72-47d4-935d-4fa777b0fc46-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4\" (UID: \"e7967b3f-3a72-47d4-935d-4fa777b0fc46\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4" Oct 07 14:07:40 crc kubenswrapper[4959]: I1007 14:07:40.212285 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e7967b3f-3a72-47d4-935d-4fa777b0fc46-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4\" (UID: \"e7967b3f-3a72-47d4-935d-4fa777b0fc46\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4" Oct 07 14:07:40 crc kubenswrapper[4959]: I1007 14:07:40.212418 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxwcb\" (UniqueName: \"kubernetes.io/projected/e7967b3f-3a72-47d4-935d-4fa777b0fc46-kube-api-access-mxwcb\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4\" (UID: \"e7967b3f-3a72-47d4-935d-4fa777b0fc46\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4" Oct 07 14:07:40 crc kubenswrapper[4959]: I1007 14:07:40.212699 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e7967b3f-3a72-47d4-935d-4fa777b0fc46-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4\" (UID: \"e7967b3f-3a72-47d4-935d-4fa777b0fc46\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4" Oct 07 14:07:40 crc kubenswrapper[4959]: I1007 14:07:40.222171 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e7967b3f-3a72-47d4-935d-4fa777b0fc46-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4\" (UID: \"e7967b3f-3a72-47d4-935d-4fa777b0fc46\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4" Oct 07 14:07:40 crc kubenswrapper[4959]: I1007 14:07:40.222584 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7967b3f-3a72-47d4-935d-4fa777b0fc46-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4\" (UID: \"e7967b3f-3a72-47d4-935d-4fa777b0fc46\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4" Oct 07 14:07:40 crc kubenswrapper[4959]: I1007 14:07:40.231874 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e7967b3f-3a72-47d4-935d-4fa777b0fc46-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4\" (UID: \"e7967b3f-3a72-47d4-935d-4fa777b0fc46\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4" Oct 07 14:07:40 crc kubenswrapper[4959]: I1007 14:07:40.233176 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxwcb\" (UniqueName: \"kubernetes.io/projected/e7967b3f-3a72-47d4-935d-4fa777b0fc46-kube-api-access-mxwcb\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4\" (UID: \"e7967b3f-3a72-47d4-935d-4fa777b0fc46\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4" Oct 07 14:07:40 crc kubenswrapper[4959]: I1007 14:07:40.338709 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4" Oct 07 14:07:40 crc kubenswrapper[4959]: I1007 14:07:40.979787 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4"] Oct 07 14:07:40 crc kubenswrapper[4959]: W1007 14:07:40.987771 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7967b3f_3a72_47d4_935d_4fa777b0fc46.slice/crio-617cc0853cf664e5b5efc77aef87b5e40f157abc4341e6a9be07ee1b25cb47a8 WatchSource:0}: Error finding container 617cc0853cf664e5b5efc77aef87b5e40f157abc4341e6a9be07ee1b25cb47a8: Status 404 returned error can't find the container with id 617cc0853cf664e5b5efc77aef87b5e40f157abc4341e6a9be07ee1b25cb47a8 Oct 07 14:07:41 crc kubenswrapper[4959]: I1007 14:07:41.955523 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4" event={"ID":"e7967b3f-3a72-47d4-935d-4fa777b0fc46","Type":"ContainerStarted","Data":"35091a54cbb649c1df69456b4413f744e19f37ada482440213a5881c959b6079"} Oct 07 14:07:41 crc kubenswrapper[4959]: I1007 14:07:41.956084 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4" event={"ID":"e7967b3f-3a72-47d4-935d-4fa777b0fc46","Type":"ContainerStarted","Data":"617cc0853cf664e5b5efc77aef87b5e40f157abc4341e6a9be07ee1b25cb47a8"} Oct 07 14:07:49 crc kubenswrapper[4959]: I1007 14:07:49.980269 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 07 14:07:50 crc kubenswrapper[4959]: I1007 14:07:50.008993 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4" podStartSLOduration=10.48757473 podStartE2EDuration="11.008974973s" podCreationTimestamp="2025-10-07 14:07:39 +0000 UTC" firstStartedPulling="2025-10-07 14:07:40.992387615 +0000 UTC m=+1343.075791940" lastFinishedPulling="2025-10-07 14:07:41.513787858 +0000 UTC m=+1343.597192183" observedRunningTime="2025-10-07 14:07:41.976393377 +0000 UTC m=+1344.059797702" watchObservedRunningTime="2025-10-07 14:07:50.008974973 +0000 UTC m=+1352.092379298" Oct 07 14:07:50 crc kubenswrapper[4959]: I1007 14:07:50.629563 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:07:50 crc kubenswrapper[4959]: I1007 14:07:50.629618 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:07:50 crc kubenswrapper[4959]: I1007 14:07:50.629656 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 14:07:50 crc kubenswrapper[4959]: I1007 14:07:50.630344 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"18f3e5766b69d97101a7543166ad26f606c175effe330b99db89cae18c7ad6ba"} pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 14:07:50 crc kubenswrapper[4959]: I1007 14:07:50.630393 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" containerID="cri-o://18f3e5766b69d97101a7543166ad26f606c175effe330b99db89cae18c7ad6ba" gracePeriod=600 Oct 07 14:07:51 crc kubenswrapper[4959]: I1007 14:07:51.028225 4959 generic.go:334] "Generic (PLEG): container finished" podID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerID="18f3e5766b69d97101a7543166ad26f606c175effe330b99db89cae18c7ad6ba" exitCode=0 Oct 07 14:07:51 crc kubenswrapper[4959]: I1007 14:07:51.028259 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 07 14:07:51 crc kubenswrapper[4959]: I1007 14:07:51.028564 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerDied","Data":"18f3e5766b69d97101a7543166ad26f606c175effe330b99db89cae18c7ad6ba"} Oct 07 14:07:51 crc kubenswrapper[4959]: I1007 14:07:51.028587 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerStarted","Data":"8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1"} Oct 07 14:07:51 crc kubenswrapper[4959]: I1007 14:07:51.028606 4959 scope.go:117] "RemoveContainer" containerID="f783f953eb3f6009a38800382e284161bac6530f08d9be67139e2b69d6c3e99d" Oct 07 14:08:27 crc kubenswrapper[4959]: I1007 14:08:27.205344 4959 scope.go:117] "RemoveContainer" containerID="bbbf601c5426d5a56d29c92700a02221d86468a30f2fbe80f508b782d4182e22" Oct 07 14:08:27 crc kubenswrapper[4959]: I1007 14:08:27.235939 4959 scope.go:117] "RemoveContainer" containerID="c0ad2f18829f8f0a34b2c312218f32a2d59a1eb1a2c1b95f5eb697e0b023c290" Oct 07 14:09:17 crc kubenswrapper[4959]: I1007 14:09:17.047261 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pf5ds"] Oct 07 14:09:17 crc kubenswrapper[4959]: I1007 14:09:17.055151 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pf5ds" Oct 07 14:09:17 crc kubenswrapper[4959]: I1007 14:09:17.082692 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pf5ds"] Oct 07 14:09:17 crc kubenswrapper[4959]: I1007 14:09:17.210736 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnw4h\" (UniqueName: \"kubernetes.io/projected/88ed62e1-eef2-4e5f-b3cf-96da2612214e-kube-api-access-mnw4h\") pod \"certified-operators-pf5ds\" (UID: \"88ed62e1-eef2-4e5f-b3cf-96da2612214e\") " pod="openshift-marketplace/certified-operators-pf5ds" Oct 07 14:09:17 crc kubenswrapper[4959]: I1007 14:09:17.210899 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88ed62e1-eef2-4e5f-b3cf-96da2612214e-catalog-content\") pod \"certified-operators-pf5ds\" (UID: \"88ed62e1-eef2-4e5f-b3cf-96da2612214e\") " pod="openshift-marketplace/certified-operators-pf5ds" Oct 07 14:09:17 crc kubenswrapper[4959]: I1007 14:09:17.210967 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88ed62e1-eef2-4e5f-b3cf-96da2612214e-utilities\") pod \"certified-operators-pf5ds\" (UID: \"88ed62e1-eef2-4e5f-b3cf-96da2612214e\") " pod="openshift-marketplace/certified-operators-pf5ds" Oct 07 14:09:17 crc kubenswrapper[4959]: I1007 14:09:17.312451 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88ed62e1-eef2-4e5f-b3cf-96da2612214e-catalog-content\") pod \"certified-operators-pf5ds\" (UID: \"88ed62e1-eef2-4e5f-b3cf-96da2612214e\") " pod="openshift-marketplace/certified-operators-pf5ds" Oct 07 14:09:17 crc kubenswrapper[4959]: I1007 14:09:17.312513 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88ed62e1-eef2-4e5f-b3cf-96da2612214e-utilities\") pod \"certified-operators-pf5ds\" (UID: \"88ed62e1-eef2-4e5f-b3cf-96da2612214e\") " pod="openshift-marketplace/certified-operators-pf5ds" Oct 07 14:09:17 crc kubenswrapper[4959]: I1007 14:09:17.312614 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnw4h\" (UniqueName: \"kubernetes.io/projected/88ed62e1-eef2-4e5f-b3cf-96da2612214e-kube-api-access-mnw4h\") pod \"certified-operators-pf5ds\" (UID: \"88ed62e1-eef2-4e5f-b3cf-96da2612214e\") " pod="openshift-marketplace/certified-operators-pf5ds" Oct 07 14:09:17 crc kubenswrapper[4959]: I1007 14:09:17.313146 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88ed62e1-eef2-4e5f-b3cf-96da2612214e-catalog-content\") pod \"certified-operators-pf5ds\" (UID: \"88ed62e1-eef2-4e5f-b3cf-96da2612214e\") " pod="openshift-marketplace/certified-operators-pf5ds" Oct 07 14:09:17 crc kubenswrapper[4959]: I1007 14:09:17.313167 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88ed62e1-eef2-4e5f-b3cf-96da2612214e-utilities\") pod \"certified-operators-pf5ds\" (UID: \"88ed62e1-eef2-4e5f-b3cf-96da2612214e\") " pod="openshift-marketplace/certified-operators-pf5ds" Oct 07 14:09:17 crc kubenswrapper[4959]: I1007 14:09:17.335822 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnw4h\" (UniqueName: \"kubernetes.io/projected/88ed62e1-eef2-4e5f-b3cf-96da2612214e-kube-api-access-mnw4h\") pod \"certified-operators-pf5ds\" (UID: \"88ed62e1-eef2-4e5f-b3cf-96da2612214e\") " pod="openshift-marketplace/certified-operators-pf5ds" Oct 07 14:09:17 crc kubenswrapper[4959]: I1007 14:09:17.379589 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pf5ds" Oct 07 14:09:17 crc kubenswrapper[4959]: I1007 14:09:17.731312 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pf5ds"] Oct 07 14:09:17 crc kubenswrapper[4959]: I1007 14:09:17.828237 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pf5ds" event={"ID":"88ed62e1-eef2-4e5f-b3cf-96da2612214e","Type":"ContainerStarted","Data":"f6f52281c4c0490d79a0045ea33ec107996889a7d016acc3b10165b824692816"} Oct 07 14:09:18 crc kubenswrapper[4959]: I1007 14:09:18.842722 4959 generic.go:334] "Generic (PLEG): container finished" podID="88ed62e1-eef2-4e5f-b3cf-96da2612214e" containerID="c16eb25ed77ca72120fe303c4232b28473ada6406e5fee55860ec66611ca7a52" exitCode=0 Oct 07 14:09:18 crc kubenswrapper[4959]: I1007 14:09:18.842806 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pf5ds" event={"ID":"88ed62e1-eef2-4e5f-b3cf-96da2612214e","Type":"ContainerDied","Data":"c16eb25ed77ca72120fe303c4232b28473ada6406e5fee55860ec66611ca7a52"} Oct 07 14:09:20 crc kubenswrapper[4959]: I1007 14:09:20.874248 4959 generic.go:334] "Generic (PLEG): container finished" podID="88ed62e1-eef2-4e5f-b3cf-96da2612214e" containerID="84e1c1c75aeb42d38dd88b045d1ae017ad0950b3d042dabf489844e8eb8e28d1" exitCode=0 Oct 07 14:09:20 crc kubenswrapper[4959]: I1007 14:09:20.874414 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pf5ds" event={"ID":"88ed62e1-eef2-4e5f-b3cf-96da2612214e","Type":"ContainerDied","Data":"84e1c1c75aeb42d38dd88b045d1ae017ad0950b3d042dabf489844e8eb8e28d1"} Oct 07 14:09:22 crc kubenswrapper[4959]: I1007 14:09:22.893561 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pf5ds" event={"ID":"88ed62e1-eef2-4e5f-b3cf-96da2612214e","Type":"ContainerStarted","Data":"1bcc43e950856acd9e326311be7b2d2dd5d146d455ab257531ba57f2d098d0cb"} Oct 07 14:09:22 crc kubenswrapper[4959]: I1007 14:09:22.921860 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pf5ds" podStartSLOduration=2.56571977 podStartE2EDuration="5.921842082s" podCreationTimestamp="2025-10-07 14:09:17 +0000 UTC" firstStartedPulling="2025-10-07 14:09:18.846078925 +0000 UTC m=+1440.929483250" lastFinishedPulling="2025-10-07 14:09:22.202201237 +0000 UTC m=+1444.285605562" observedRunningTime="2025-10-07 14:09:22.917197476 +0000 UTC m=+1445.000601801" watchObservedRunningTime="2025-10-07 14:09:22.921842082 +0000 UTC m=+1445.005246407" Oct 07 14:09:27 crc kubenswrapper[4959]: I1007 14:09:27.329344 4959 scope.go:117] "RemoveContainer" containerID="2718b09b1af576bdba5f8d95924e279615a246feeba4cd05f9a0e23a587642f1" Oct 07 14:09:27 crc kubenswrapper[4959]: I1007 14:09:27.380408 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pf5ds" Oct 07 14:09:27 crc kubenswrapper[4959]: I1007 14:09:27.404129 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pf5ds" Oct 07 14:09:27 crc kubenswrapper[4959]: I1007 14:09:27.447435 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pf5ds" Oct 07 14:09:27 crc kubenswrapper[4959]: I1007 14:09:27.983408 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pf5ds" Oct 07 14:09:28 crc kubenswrapper[4959]: I1007 14:09:28.036022 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pf5ds"] Oct 07 14:09:29 crc kubenswrapper[4959]: I1007 14:09:29.954140 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pf5ds" podUID="88ed62e1-eef2-4e5f-b3cf-96da2612214e" containerName="registry-server" containerID="cri-o://1bcc43e950856acd9e326311be7b2d2dd5d146d455ab257531ba57f2d098d0cb" gracePeriod=2 Oct 07 14:09:30 crc kubenswrapper[4959]: I1007 14:09:30.963083 4959 generic.go:334] "Generic (PLEG): container finished" podID="88ed62e1-eef2-4e5f-b3cf-96da2612214e" containerID="1bcc43e950856acd9e326311be7b2d2dd5d146d455ab257531ba57f2d098d0cb" exitCode=0 Oct 07 14:09:30 crc kubenswrapper[4959]: I1007 14:09:30.963127 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pf5ds" event={"ID":"88ed62e1-eef2-4e5f-b3cf-96da2612214e","Type":"ContainerDied","Data":"1bcc43e950856acd9e326311be7b2d2dd5d146d455ab257531ba57f2d098d0cb"} Oct 07 14:09:31 crc kubenswrapper[4959]: I1007 14:09:31.172123 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pf5ds" Oct 07 14:09:31 crc kubenswrapper[4959]: I1007 14:09:31.298209 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88ed62e1-eef2-4e5f-b3cf-96da2612214e-utilities\") pod \"88ed62e1-eef2-4e5f-b3cf-96da2612214e\" (UID: \"88ed62e1-eef2-4e5f-b3cf-96da2612214e\") " Oct 07 14:09:31 crc kubenswrapper[4959]: I1007 14:09:31.298318 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnw4h\" (UniqueName: \"kubernetes.io/projected/88ed62e1-eef2-4e5f-b3cf-96da2612214e-kube-api-access-mnw4h\") pod \"88ed62e1-eef2-4e5f-b3cf-96da2612214e\" (UID: \"88ed62e1-eef2-4e5f-b3cf-96da2612214e\") " Oct 07 14:09:31 crc kubenswrapper[4959]: I1007 14:09:31.298383 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88ed62e1-eef2-4e5f-b3cf-96da2612214e-catalog-content\") pod \"88ed62e1-eef2-4e5f-b3cf-96da2612214e\" (UID: \"88ed62e1-eef2-4e5f-b3cf-96da2612214e\") " Oct 07 14:09:31 crc kubenswrapper[4959]: I1007 14:09:31.299811 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88ed62e1-eef2-4e5f-b3cf-96da2612214e-utilities" (OuterVolumeSpecName: "utilities") pod "88ed62e1-eef2-4e5f-b3cf-96da2612214e" (UID: "88ed62e1-eef2-4e5f-b3cf-96da2612214e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:09:31 crc kubenswrapper[4959]: I1007 14:09:31.305889 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88ed62e1-eef2-4e5f-b3cf-96da2612214e-kube-api-access-mnw4h" (OuterVolumeSpecName: "kube-api-access-mnw4h") pod "88ed62e1-eef2-4e5f-b3cf-96da2612214e" (UID: "88ed62e1-eef2-4e5f-b3cf-96da2612214e"). InnerVolumeSpecName "kube-api-access-mnw4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:09:31 crc kubenswrapper[4959]: I1007 14:09:31.357990 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88ed62e1-eef2-4e5f-b3cf-96da2612214e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "88ed62e1-eef2-4e5f-b3cf-96da2612214e" (UID: "88ed62e1-eef2-4e5f-b3cf-96da2612214e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:09:31 crc kubenswrapper[4959]: I1007 14:09:31.399992 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88ed62e1-eef2-4e5f-b3cf-96da2612214e-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 14:09:31 crc kubenswrapper[4959]: I1007 14:09:31.400030 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnw4h\" (UniqueName: \"kubernetes.io/projected/88ed62e1-eef2-4e5f-b3cf-96da2612214e-kube-api-access-mnw4h\") on node \"crc\" DevicePath \"\"" Oct 07 14:09:31 crc kubenswrapper[4959]: I1007 14:09:31.400041 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88ed62e1-eef2-4e5f-b3cf-96da2612214e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 14:09:31 crc kubenswrapper[4959]: I1007 14:09:31.979397 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pf5ds" event={"ID":"88ed62e1-eef2-4e5f-b3cf-96da2612214e","Type":"ContainerDied","Data":"f6f52281c4c0490d79a0045ea33ec107996889a7d016acc3b10165b824692816"} Oct 07 14:09:31 crc kubenswrapper[4959]: I1007 14:09:31.979474 4959 scope.go:117] "RemoveContainer" containerID="1bcc43e950856acd9e326311be7b2d2dd5d146d455ab257531ba57f2d098d0cb" Oct 07 14:09:31 crc kubenswrapper[4959]: I1007 14:09:31.979505 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pf5ds" Oct 07 14:09:32 crc kubenswrapper[4959]: I1007 14:09:32.023734 4959 scope.go:117] "RemoveContainer" containerID="84e1c1c75aeb42d38dd88b045d1ae017ad0950b3d042dabf489844e8eb8e28d1" Oct 07 14:09:32 crc kubenswrapper[4959]: I1007 14:09:32.026058 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pf5ds"] Oct 07 14:09:32 crc kubenswrapper[4959]: I1007 14:09:32.034322 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pf5ds"] Oct 07 14:09:32 crc kubenswrapper[4959]: I1007 14:09:32.055748 4959 scope.go:117] "RemoveContainer" containerID="c16eb25ed77ca72120fe303c4232b28473ada6406e5fee55860ec66611ca7a52" Oct 07 14:09:32 crc kubenswrapper[4959]: I1007 14:09:32.665904 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88ed62e1-eef2-4e5f-b3cf-96da2612214e" path="/var/lib/kubelet/pods/88ed62e1-eef2-4e5f-b3cf-96da2612214e/volumes" Oct 07 14:09:46 crc kubenswrapper[4959]: I1007 14:09:46.270217 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pz5dh"] Oct 07 14:09:46 crc kubenswrapper[4959]: E1007 14:09:46.271246 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88ed62e1-eef2-4e5f-b3cf-96da2612214e" containerName="extract-content" Oct 07 14:09:46 crc kubenswrapper[4959]: I1007 14:09:46.271264 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="88ed62e1-eef2-4e5f-b3cf-96da2612214e" containerName="extract-content" Oct 07 14:09:46 crc kubenswrapper[4959]: E1007 14:09:46.271288 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88ed62e1-eef2-4e5f-b3cf-96da2612214e" containerName="registry-server" Oct 07 14:09:46 crc kubenswrapper[4959]: I1007 14:09:46.271297 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="88ed62e1-eef2-4e5f-b3cf-96da2612214e" containerName="registry-server" Oct 07 14:09:46 crc kubenswrapper[4959]: E1007 14:09:46.271325 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88ed62e1-eef2-4e5f-b3cf-96da2612214e" containerName="extract-utilities" Oct 07 14:09:46 crc kubenswrapper[4959]: I1007 14:09:46.271334 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="88ed62e1-eef2-4e5f-b3cf-96da2612214e" containerName="extract-utilities" Oct 07 14:09:46 crc kubenswrapper[4959]: I1007 14:09:46.271561 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="88ed62e1-eef2-4e5f-b3cf-96da2612214e" containerName="registry-server" Oct 07 14:09:46 crc kubenswrapper[4959]: I1007 14:09:46.273071 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pz5dh" Oct 07 14:09:46 crc kubenswrapper[4959]: I1007 14:09:46.343092 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pz5dh"] Oct 07 14:09:46 crc kubenswrapper[4959]: I1007 14:09:46.474915 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec297f63-be6e-44c2-9ebb-8acdecb3d31f-utilities\") pod \"redhat-marketplace-pz5dh\" (UID: \"ec297f63-be6e-44c2-9ebb-8acdecb3d31f\") " pod="openshift-marketplace/redhat-marketplace-pz5dh" Oct 07 14:09:46 crc kubenswrapper[4959]: I1007 14:09:46.475006 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec297f63-be6e-44c2-9ebb-8acdecb3d31f-catalog-content\") pod \"redhat-marketplace-pz5dh\" (UID: \"ec297f63-be6e-44c2-9ebb-8acdecb3d31f\") " pod="openshift-marketplace/redhat-marketplace-pz5dh" Oct 07 14:09:46 crc kubenswrapper[4959]: I1007 14:09:46.475357 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spbdr\" (UniqueName: \"kubernetes.io/projected/ec297f63-be6e-44c2-9ebb-8acdecb3d31f-kube-api-access-spbdr\") pod \"redhat-marketplace-pz5dh\" (UID: \"ec297f63-be6e-44c2-9ebb-8acdecb3d31f\") " pod="openshift-marketplace/redhat-marketplace-pz5dh" Oct 07 14:09:46 crc kubenswrapper[4959]: I1007 14:09:46.577043 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spbdr\" (UniqueName: \"kubernetes.io/projected/ec297f63-be6e-44c2-9ebb-8acdecb3d31f-kube-api-access-spbdr\") pod \"redhat-marketplace-pz5dh\" (UID: \"ec297f63-be6e-44c2-9ebb-8acdecb3d31f\") " pod="openshift-marketplace/redhat-marketplace-pz5dh" Oct 07 14:09:46 crc kubenswrapper[4959]: I1007 14:09:46.577118 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec297f63-be6e-44c2-9ebb-8acdecb3d31f-utilities\") pod \"redhat-marketplace-pz5dh\" (UID: \"ec297f63-be6e-44c2-9ebb-8acdecb3d31f\") " pod="openshift-marketplace/redhat-marketplace-pz5dh" Oct 07 14:09:46 crc kubenswrapper[4959]: I1007 14:09:46.577177 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec297f63-be6e-44c2-9ebb-8acdecb3d31f-catalog-content\") pod \"redhat-marketplace-pz5dh\" (UID: \"ec297f63-be6e-44c2-9ebb-8acdecb3d31f\") " pod="openshift-marketplace/redhat-marketplace-pz5dh" Oct 07 14:09:46 crc kubenswrapper[4959]: I1007 14:09:46.577622 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec297f63-be6e-44c2-9ebb-8acdecb3d31f-utilities\") pod \"redhat-marketplace-pz5dh\" (UID: \"ec297f63-be6e-44c2-9ebb-8acdecb3d31f\") " pod="openshift-marketplace/redhat-marketplace-pz5dh" Oct 07 14:09:46 crc kubenswrapper[4959]: I1007 14:09:46.577773 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec297f63-be6e-44c2-9ebb-8acdecb3d31f-catalog-content\") pod \"redhat-marketplace-pz5dh\" (UID: \"ec297f63-be6e-44c2-9ebb-8acdecb3d31f\") " pod="openshift-marketplace/redhat-marketplace-pz5dh" Oct 07 14:09:46 crc kubenswrapper[4959]: I1007 14:09:46.601135 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spbdr\" (UniqueName: \"kubernetes.io/projected/ec297f63-be6e-44c2-9ebb-8acdecb3d31f-kube-api-access-spbdr\") pod \"redhat-marketplace-pz5dh\" (UID: \"ec297f63-be6e-44c2-9ebb-8acdecb3d31f\") " pod="openshift-marketplace/redhat-marketplace-pz5dh" Oct 07 14:09:46 crc kubenswrapper[4959]: I1007 14:09:46.646381 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pz5dh" Oct 07 14:09:47 crc kubenswrapper[4959]: I1007 14:09:47.092205 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pz5dh"] Oct 07 14:09:48 crc kubenswrapper[4959]: I1007 14:09:48.124393 4959 generic.go:334] "Generic (PLEG): container finished" podID="ec297f63-be6e-44c2-9ebb-8acdecb3d31f" containerID="d10c8223c1a39325bd8397025261696e3b3c3653e333d4003b862da936b2effd" exitCode=0 Oct 07 14:09:48 crc kubenswrapper[4959]: I1007 14:09:48.124711 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pz5dh" event={"ID":"ec297f63-be6e-44c2-9ebb-8acdecb3d31f","Type":"ContainerDied","Data":"d10c8223c1a39325bd8397025261696e3b3c3653e333d4003b862da936b2effd"} Oct 07 14:09:48 crc kubenswrapper[4959]: I1007 14:09:48.124752 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pz5dh" event={"ID":"ec297f63-be6e-44c2-9ebb-8acdecb3d31f","Type":"ContainerStarted","Data":"667bce713a9cebffb1f22f4de9cb529e8d68913c0b56ff308e2670351e00fcbe"} Oct 07 14:09:49 crc kubenswrapper[4959]: I1007 14:09:49.145080 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pz5dh" event={"ID":"ec297f63-be6e-44c2-9ebb-8acdecb3d31f","Type":"ContainerStarted","Data":"e8775975d8376042e7a05e07e968a85cd219c6da9331b1db193b0ad4cee55b78"} Oct 07 14:09:50 crc kubenswrapper[4959]: I1007 14:09:50.155507 4959 generic.go:334] "Generic (PLEG): container finished" podID="ec297f63-be6e-44c2-9ebb-8acdecb3d31f" containerID="e8775975d8376042e7a05e07e968a85cd219c6da9331b1db193b0ad4cee55b78" exitCode=0 Oct 07 14:09:50 crc kubenswrapper[4959]: I1007 14:09:50.155604 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pz5dh" event={"ID":"ec297f63-be6e-44c2-9ebb-8acdecb3d31f","Type":"ContainerDied","Data":"e8775975d8376042e7a05e07e968a85cd219c6da9331b1db193b0ad4cee55b78"} Oct 07 14:09:50 crc kubenswrapper[4959]: I1007 14:09:50.156321 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pz5dh" event={"ID":"ec297f63-be6e-44c2-9ebb-8acdecb3d31f","Type":"ContainerStarted","Data":"69c9db247d8173898a94ca1b5345ddc1484228d0ddbb2d33e3ca9a4d52bc9db7"} Oct 07 14:09:50 crc kubenswrapper[4959]: I1007 14:09:50.178882 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pz5dh" podStartSLOduration=2.451861083 podStartE2EDuration="4.178858913s" podCreationTimestamp="2025-10-07 14:09:46 +0000 UTC" firstStartedPulling="2025-10-07 14:09:48.126917995 +0000 UTC m=+1470.210322330" lastFinishedPulling="2025-10-07 14:09:49.853915835 +0000 UTC m=+1471.937320160" observedRunningTime="2025-10-07 14:09:50.175266494 +0000 UTC m=+1472.258670849" watchObservedRunningTime="2025-10-07 14:09:50.178858913 +0000 UTC m=+1472.262263238" Oct 07 14:09:50 crc kubenswrapper[4959]: I1007 14:09:50.630338 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:09:50 crc kubenswrapper[4959]: I1007 14:09:50.630398 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:09:56 crc kubenswrapper[4959]: I1007 14:09:56.646509 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pz5dh" Oct 07 14:09:56 crc kubenswrapper[4959]: I1007 14:09:56.647043 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pz5dh" Oct 07 14:09:56 crc kubenswrapper[4959]: I1007 14:09:56.696012 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pz5dh" Oct 07 14:09:57 crc kubenswrapper[4959]: I1007 14:09:57.268048 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pz5dh" Oct 07 14:09:57 crc kubenswrapper[4959]: I1007 14:09:57.320068 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pz5dh"] Oct 07 14:09:59 crc kubenswrapper[4959]: I1007 14:09:59.232709 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pz5dh" podUID="ec297f63-be6e-44c2-9ebb-8acdecb3d31f" containerName="registry-server" containerID="cri-o://69c9db247d8173898a94ca1b5345ddc1484228d0ddbb2d33e3ca9a4d52bc9db7" gracePeriod=2 Oct 07 14:09:59 crc kubenswrapper[4959]: I1007 14:09:59.707547 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pz5dh" Oct 07 14:09:59 crc kubenswrapper[4959]: I1007 14:09:59.857574 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spbdr\" (UniqueName: \"kubernetes.io/projected/ec297f63-be6e-44c2-9ebb-8acdecb3d31f-kube-api-access-spbdr\") pod \"ec297f63-be6e-44c2-9ebb-8acdecb3d31f\" (UID: \"ec297f63-be6e-44c2-9ebb-8acdecb3d31f\") " Oct 07 14:09:59 crc kubenswrapper[4959]: I1007 14:09:59.857748 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec297f63-be6e-44c2-9ebb-8acdecb3d31f-utilities\") pod \"ec297f63-be6e-44c2-9ebb-8acdecb3d31f\" (UID: \"ec297f63-be6e-44c2-9ebb-8acdecb3d31f\") " Oct 07 14:09:59 crc kubenswrapper[4959]: I1007 14:09:59.857831 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec297f63-be6e-44c2-9ebb-8acdecb3d31f-catalog-content\") pod \"ec297f63-be6e-44c2-9ebb-8acdecb3d31f\" (UID: \"ec297f63-be6e-44c2-9ebb-8acdecb3d31f\") " Oct 07 14:09:59 crc kubenswrapper[4959]: I1007 14:09:59.858920 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec297f63-be6e-44c2-9ebb-8acdecb3d31f-utilities" (OuterVolumeSpecName: "utilities") pod "ec297f63-be6e-44c2-9ebb-8acdecb3d31f" (UID: "ec297f63-be6e-44c2-9ebb-8acdecb3d31f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:09:59 crc kubenswrapper[4959]: I1007 14:09:59.863835 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec297f63-be6e-44c2-9ebb-8acdecb3d31f-kube-api-access-spbdr" (OuterVolumeSpecName: "kube-api-access-spbdr") pod "ec297f63-be6e-44c2-9ebb-8acdecb3d31f" (UID: "ec297f63-be6e-44c2-9ebb-8acdecb3d31f"). InnerVolumeSpecName "kube-api-access-spbdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:09:59 crc kubenswrapper[4959]: I1007 14:09:59.871134 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec297f63-be6e-44c2-9ebb-8acdecb3d31f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ec297f63-be6e-44c2-9ebb-8acdecb3d31f" (UID: "ec297f63-be6e-44c2-9ebb-8acdecb3d31f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:09:59 crc kubenswrapper[4959]: I1007 14:09:59.960410 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec297f63-be6e-44c2-9ebb-8acdecb3d31f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 14:09:59 crc kubenswrapper[4959]: I1007 14:09:59.960466 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spbdr\" (UniqueName: \"kubernetes.io/projected/ec297f63-be6e-44c2-9ebb-8acdecb3d31f-kube-api-access-spbdr\") on node \"crc\" DevicePath \"\"" Oct 07 14:09:59 crc kubenswrapper[4959]: I1007 14:09:59.960481 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec297f63-be6e-44c2-9ebb-8acdecb3d31f-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 14:10:00 crc kubenswrapper[4959]: I1007 14:10:00.243447 4959 generic.go:334] "Generic (PLEG): container finished" podID="ec297f63-be6e-44c2-9ebb-8acdecb3d31f" containerID="69c9db247d8173898a94ca1b5345ddc1484228d0ddbb2d33e3ca9a4d52bc9db7" exitCode=0 Oct 07 14:10:00 crc kubenswrapper[4959]: I1007 14:10:00.243509 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pz5dh" event={"ID":"ec297f63-be6e-44c2-9ebb-8acdecb3d31f","Type":"ContainerDied","Data":"69c9db247d8173898a94ca1b5345ddc1484228d0ddbb2d33e3ca9a4d52bc9db7"} Oct 07 14:10:00 crc kubenswrapper[4959]: I1007 14:10:00.243560 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pz5dh" event={"ID":"ec297f63-be6e-44c2-9ebb-8acdecb3d31f","Type":"ContainerDied","Data":"667bce713a9cebffb1f22f4de9cb529e8d68913c0b56ff308e2670351e00fcbe"} Oct 07 14:10:00 crc kubenswrapper[4959]: I1007 14:10:00.243583 4959 scope.go:117] "RemoveContainer" containerID="69c9db247d8173898a94ca1b5345ddc1484228d0ddbb2d33e3ca9a4d52bc9db7" Oct 07 14:10:00 crc kubenswrapper[4959]: I1007 14:10:00.243583 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pz5dh" Oct 07 14:10:00 crc kubenswrapper[4959]: I1007 14:10:00.271508 4959 scope.go:117] "RemoveContainer" containerID="e8775975d8376042e7a05e07e968a85cd219c6da9331b1db193b0ad4cee55b78" Oct 07 14:10:00 crc kubenswrapper[4959]: I1007 14:10:00.275256 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pz5dh"] Oct 07 14:10:00 crc kubenswrapper[4959]: I1007 14:10:00.282576 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pz5dh"] Oct 07 14:10:00 crc kubenswrapper[4959]: I1007 14:10:00.296884 4959 scope.go:117] "RemoveContainer" containerID="d10c8223c1a39325bd8397025261696e3b3c3653e333d4003b862da936b2effd" Oct 07 14:10:00 crc kubenswrapper[4959]: I1007 14:10:00.332608 4959 scope.go:117] "RemoveContainer" containerID="69c9db247d8173898a94ca1b5345ddc1484228d0ddbb2d33e3ca9a4d52bc9db7" Oct 07 14:10:00 crc kubenswrapper[4959]: E1007 14:10:00.333050 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69c9db247d8173898a94ca1b5345ddc1484228d0ddbb2d33e3ca9a4d52bc9db7\": container with ID starting with 69c9db247d8173898a94ca1b5345ddc1484228d0ddbb2d33e3ca9a4d52bc9db7 not found: ID does not exist" containerID="69c9db247d8173898a94ca1b5345ddc1484228d0ddbb2d33e3ca9a4d52bc9db7" Oct 07 14:10:00 crc kubenswrapper[4959]: I1007 14:10:00.333085 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69c9db247d8173898a94ca1b5345ddc1484228d0ddbb2d33e3ca9a4d52bc9db7"} err="failed to get container status \"69c9db247d8173898a94ca1b5345ddc1484228d0ddbb2d33e3ca9a4d52bc9db7\": rpc error: code = NotFound desc = could not find container \"69c9db247d8173898a94ca1b5345ddc1484228d0ddbb2d33e3ca9a4d52bc9db7\": container with ID starting with 69c9db247d8173898a94ca1b5345ddc1484228d0ddbb2d33e3ca9a4d52bc9db7 not found: ID does not exist" Oct 07 14:10:00 crc kubenswrapper[4959]: I1007 14:10:00.333129 4959 scope.go:117] "RemoveContainer" containerID="e8775975d8376042e7a05e07e968a85cd219c6da9331b1db193b0ad4cee55b78" Oct 07 14:10:00 crc kubenswrapper[4959]: E1007 14:10:00.333366 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8775975d8376042e7a05e07e968a85cd219c6da9331b1db193b0ad4cee55b78\": container with ID starting with e8775975d8376042e7a05e07e968a85cd219c6da9331b1db193b0ad4cee55b78 not found: ID does not exist" containerID="e8775975d8376042e7a05e07e968a85cd219c6da9331b1db193b0ad4cee55b78" Oct 07 14:10:00 crc kubenswrapper[4959]: I1007 14:10:00.333391 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8775975d8376042e7a05e07e968a85cd219c6da9331b1db193b0ad4cee55b78"} err="failed to get container status \"e8775975d8376042e7a05e07e968a85cd219c6da9331b1db193b0ad4cee55b78\": rpc error: code = NotFound desc = could not find container \"e8775975d8376042e7a05e07e968a85cd219c6da9331b1db193b0ad4cee55b78\": container with ID starting with e8775975d8376042e7a05e07e968a85cd219c6da9331b1db193b0ad4cee55b78 not found: ID does not exist" Oct 07 14:10:00 crc kubenswrapper[4959]: I1007 14:10:00.333405 4959 scope.go:117] "RemoveContainer" containerID="d10c8223c1a39325bd8397025261696e3b3c3653e333d4003b862da936b2effd" Oct 07 14:10:00 crc kubenswrapper[4959]: E1007 14:10:00.333643 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d10c8223c1a39325bd8397025261696e3b3c3653e333d4003b862da936b2effd\": container with ID starting with d10c8223c1a39325bd8397025261696e3b3c3653e333d4003b862da936b2effd not found: ID does not exist" containerID="d10c8223c1a39325bd8397025261696e3b3c3653e333d4003b862da936b2effd" Oct 07 14:10:00 crc kubenswrapper[4959]: I1007 14:10:00.333671 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d10c8223c1a39325bd8397025261696e3b3c3653e333d4003b862da936b2effd"} err="failed to get container status \"d10c8223c1a39325bd8397025261696e3b3c3653e333d4003b862da936b2effd\": rpc error: code = NotFound desc = could not find container \"d10c8223c1a39325bd8397025261696e3b3c3653e333d4003b862da936b2effd\": container with ID starting with d10c8223c1a39325bd8397025261696e3b3c3653e333d4003b862da936b2effd not found: ID does not exist" Oct 07 14:10:00 crc kubenswrapper[4959]: I1007 14:10:00.662791 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec297f63-be6e-44c2-9ebb-8acdecb3d31f" path="/var/lib/kubelet/pods/ec297f63-be6e-44c2-9ebb-8acdecb3d31f/volumes" Oct 07 14:10:20 crc kubenswrapper[4959]: I1007 14:10:20.630552 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:10:20 crc kubenswrapper[4959]: I1007 14:10:20.631631 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:10:40 crc kubenswrapper[4959]: I1007 14:10:40.558011 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7b2n7"] Oct 07 14:10:40 crc kubenswrapper[4959]: E1007 14:10:40.559515 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec297f63-be6e-44c2-9ebb-8acdecb3d31f" containerName="registry-server" Oct 07 14:10:40 crc kubenswrapper[4959]: I1007 14:10:40.559529 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec297f63-be6e-44c2-9ebb-8acdecb3d31f" containerName="registry-server" Oct 07 14:10:40 crc kubenswrapper[4959]: E1007 14:10:40.559542 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec297f63-be6e-44c2-9ebb-8acdecb3d31f" containerName="extract-content" Oct 07 14:10:40 crc kubenswrapper[4959]: I1007 14:10:40.559548 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec297f63-be6e-44c2-9ebb-8acdecb3d31f" containerName="extract-content" Oct 07 14:10:40 crc kubenswrapper[4959]: E1007 14:10:40.559568 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec297f63-be6e-44c2-9ebb-8acdecb3d31f" containerName="extract-utilities" Oct 07 14:10:40 crc kubenswrapper[4959]: I1007 14:10:40.559574 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec297f63-be6e-44c2-9ebb-8acdecb3d31f" containerName="extract-utilities" Oct 07 14:10:40 crc kubenswrapper[4959]: I1007 14:10:40.559754 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec297f63-be6e-44c2-9ebb-8acdecb3d31f" containerName="registry-server" Oct 07 14:10:40 crc kubenswrapper[4959]: I1007 14:10:40.561057 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7b2n7" Oct 07 14:10:40 crc kubenswrapper[4959]: I1007 14:10:40.569329 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7b2n7"] Oct 07 14:10:40 crc kubenswrapper[4959]: I1007 14:10:40.733056 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/964cde77-c6b5-4a61-a739-5d6cd089ddad-catalog-content\") pod \"redhat-operators-7b2n7\" (UID: \"964cde77-c6b5-4a61-a739-5d6cd089ddad\") " pod="openshift-marketplace/redhat-operators-7b2n7" Oct 07 14:10:40 crc kubenswrapper[4959]: I1007 14:10:40.733231 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/964cde77-c6b5-4a61-a739-5d6cd089ddad-utilities\") pod \"redhat-operators-7b2n7\" (UID: \"964cde77-c6b5-4a61-a739-5d6cd089ddad\") " pod="openshift-marketplace/redhat-operators-7b2n7" Oct 07 14:10:40 crc kubenswrapper[4959]: I1007 14:10:40.733292 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gc4f8\" (UniqueName: \"kubernetes.io/projected/964cde77-c6b5-4a61-a739-5d6cd089ddad-kube-api-access-gc4f8\") pod \"redhat-operators-7b2n7\" (UID: \"964cde77-c6b5-4a61-a739-5d6cd089ddad\") " pod="openshift-marketplace/redhat-operators-7b2n7" Oct 07 14:10:40 crc kubenswrapper[4959]: I1007 14:10:40.835361 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gc4f8\" (UniqueName: \"kubernetes.io/projected/964cde77-c6b5-4a61-a739-5d6cd089ddad-kube-api-access-gc4f8\") pod \"redhat-operators-7b2n7\" (UID: \"964cde77-c6b5-4a61-a739-5d6cd089ddad\") " pod="openshift-marketplace/redhat-operators-7b2n7" Oct 07 14:10:40 crc kubenswrapper[4959]: I1007 14:10:40.835480 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/964cde77-c6b5-4a61-a739-5d6cd089ddad-catalog-content\") pod \"redhat-operators-7b2n7\" (UID: \"964cde77-c6b5-4a61-a739-5d6cd089ddad\") " pod="openshift-marketplace/redhat-operators-7b2n7" Oct 07 14:10:40 crc kubenswrapper[4959]: I1007 14:10:40.836215 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/964cde77-c6b5-4a61-a739-5d6cd089ddad-catalog-content\") pod \"redhat-operators-7b2n7\" (UID: \"964cde77-c6b5-4a61-a739-5d6cd089ddad\") " pod="openshift-marketplace/redhat-operators-7b2n7" Oct 07 14:10:40 crc kubenswrapper[4959]: I1007 14:10:40.838991 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/964cde77-c6b5-4a61-a739-5d6cd089ddad-utilities\") pod \"redhat-operators-7b2n7\" (UID: \"964cde77-c6b5-4a61-a739-5d6cd089ddad\") " pod="openshift-marketplace/redhat-operators-7b2n7" Oct 07 14:10:40 crc kubenswrapper[4959]: I1007 14:10:40.839353 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/964cde77-c6b5-4a61-a739-5d6cd089ddad-utilities\") pod \"redhat-operators-7b2n7\" (UID: \"964cde77-c6b5-4a61-a739-5d6cd089ddad\") " pod="openshift-marketplace/redhat-operators-7b2n7" Oct 07 14:10:40 crc kubenswrapper[4959]: I1007 14:10:40.856349 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gc4f8\" (UniqueName: \"kubernetes.io/projected/964cde77-c6b5-4a61-a739-5d6cd089ddad-kube-api-access-gc4f8\") pod \"redhat-operators-7b2n7\" (UID: \"964cde77-c6b5-4a61-a739-5d6cd089ddad\") " pod="openshift-marketplace/redhat-operators-7b2n7" Oct 07 14:10:40 crc kubenswrapper[4959]: I1007 14:10:40.893976 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7b2n7" Oct 07 14:10:41 crc kubenswrapper[4959]: I1007 14:10:41.349137 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7b2n7"] Oct 07 14:10:41 crc kubenswrapper[4959]: I1007 14:10:41.626154 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7b2n7" event={"ID":"964cde77-c6b5-4a61-a739-5d6cd089ddad","Type":"ContainerStarted","Data":"0506199fc3d409aa53fb96796b07ffe1860c33d19c30415df2914fe562ba555e"} Oct 07 14:10:42 crc kubenswrapper[4959]: I1007 14:10:42.634785 4959 generic.go:334] "Generic (PLEG): container finished" podID="964cde77-c6b5-4a61-a739-5d6cd089ddad" containerID="baa35a208bf268fdc594266b821d2223bed497cc913d642a1504ce79b97c3a7b" exitCode=0 Oct 07 14:10:42 crc kubenswrapper[4959]: I1007 14:10:42.634831 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7b2n7" event={"ID":"964cde77-c6b5-4a61-a739-5d6cd089ddad","Type":"ContainerDied","Data":"baa35a208bf268fdc594266b821d2223bed497cc913d642a1504ce79b97c3a7b"} Oct 07 14:10:44 crc kubenswrapper[4959]: I1007 14:10:44.658720 4959 generic.go:334] "Generic (PLEG): container finished" podID="964cde77-c6b5-4a61-a739-5d6cd089ddad" containerID="7895595e3c229042606a66ca2da681f0c45c7a9cce4918fb6da2d7ba37a01f5d" exitCode=0 Oct 07 14:10:44 crc kubenswrapper[4959]: I1007 14:10:44.663506 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7b2n7" event={"ID":"964cde77-c6b5-4a61-a739-5d6cd089ddad","Type":"ContainerDied","Data":"7895595e3c229042606a66ca2da681f0c45c7a9cce4918fb6da2d7ba37a01f5d"} Oct 07 14:10:46 crc kubenswrapper[4959]: I1007 14:10:46.675447 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7b2n7" event={"ID":"964cde77-c6b5-4a61-a739-5d6cd089ddad","Type":"ContainerStarted","Data":"6e20f8686125a7697fbc5a68dcec1e5027ffa6b5d6a4b979fcd8245f94982da7"} Oct 07 14:10:46 crc kubenswrapper[4959]: I1007 14:10:46.704576 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7b2n7" podStartSLOduration=4.155308359 podStartE2EDuration="6.704552532s" podCreationTimestamp="2025-10-07 14:10:40 +0000 UTC" firstStartedPulling="2025-10-07 14:10:42.637336219 +0000 UTC m=+1524.720740544" lastFinishedPulling="2025-10-07 14:10:45.186580392 +0000 UTC m=+1527.269984717" observedRunningTime="2025-10-07 14:10:46.694207804 +0000 UTC m=+1528.777612159" watchObservedRunningTime="2025-10-07 14:10:46.704552532 +0000 UTC m=+1528.787956867" Oct 07 14:10:50 crc kubenswrapper[4959]: I1007 14:10:50.630561 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:10:50 crc kubenswrapper[4959]: I1007 14:10:50.631176 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:10:50 crc kubenswrapper[4959]: I1007 14:10:50.631226 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 14:10:50 crc kubenswrapper[4959]: I1007 14:10:50.631874 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1"} pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 14:10:50 crc kubenswrapper[4959]: I1007 14:10:50.631918 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" containerID="cri-o://8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1" gracePeriod=600 Oct 07 14:10:50 crc kubenswrapper[4959]: E1007 14:10:50.791921 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:10:50 crc kubenswrapper[4959]: I1007 14:10:50.895630 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7b2n7" Oct 07 14:10:50 crc kubenswrapper[4959]: I1007 14:10:50.895676 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7b2n7" Oct 07 14:10:50 crc kubenswrapper[4959]: I1007 14:10:50.943780 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7b2n7" Oct 07 14:10:51 crc kubenswrapper[4959]: I1007 14:10:51.715963 4959 generic.go:334] "Generic (PLEG): container finished" podID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerID="8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1" exitCode=0 Oct 07 14:10:51 crc kubenswrapper[4959]: I1007 14:10:51.716030 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerDied","Data":"8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1"} Oct 07 14:10:51 crc kubenswrapper[4959]: I1007 14:10:51.716495 4959 scope.go:117] "RemoveContainer" containerID="18f3e5766b69d97101a7543166ad26f606c175effe330b99db89cae18c7ad6ba" Oct 07 14:10:51 crc kubenswrapper[4959]: I1007 14:10:51.717123 4959 scope.go:117] "RemoveContainer" containerID="8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1" Oct 07 14:10:51 crc kubenswrapper[4959]: E1007 14:10:51.717474 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:10:51 crc kubenswrapper[4959]: I1007 14:10:51.792544 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7b2n7" Oct 07 14:10:51 crc kubenswrapper[4959]: I1007 14:10:51.848491 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7b2n7"] Oct 07 14:10:53 crc kubenswrapper[4959]: I1007 14:10:53.734216 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7b2n7" podUID="964cde77-c6b5-4a61-a739-5d6cd089ddad" containerName="registry-server" containerID="cri-o://6e20f8686125a7697fbc5a68dcec1e5027ffa6b5d6a4b979fcd8245f94982da7" gracePeriod=2 Oct 07 14:10:54 crc kubenswrapper[4959]: I1007 14:10:54.186602 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7b2n7" Oct 07 14:10:54 crc kubenswrapper[4959]: I1007 14:10:54.313210 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/964cde77-c6b5-4a61-a739-5d6cd089ddad-utilities\") pod \"964cde77-c6b5-4a61-a739-5d6cd089ddad\" (UID: \"964cde77-c6b5-4a61-a739-5d6cd089ddad\") " Oct 07 14:10:54 crc kubenswrapper[4959]: I1007 14:10:54.313455 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gc4f8\" (UniqueName: \"kubernetes.io/projected/964cde77-c6b5-4a61-a739-5d6cd089ddad-kube-api-access-gc4f8\") pod \"964cde77-c6b5-4a61-a739-5d6cd089ddad\" (UID: \"964cde77-c6b5-4a61-a739-5d6cd089ddad\") " Oct 07 14:10:54 crc kubenswrapper[4959]: I1007 14:10:54.313526 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/964cde77-c6b5-4a61-a739-5d6cd089ddad-catalog-content\") pod \"964cde77-c6b5-4a61-a739-5d6cd089ddad\" (UID: \"964cde77-c6b5-4a61-a739-5d6cd089ddad\") " Oct 07 14:10:54 crc kubenswrapper[4959]: I1007 14:10:54.315476 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/964cde77-c6b5-4a61-a739-5d6cd089ddad-utilities" (OuterVolumeSpecName: "utilities") pod "964cde77-c6b5-4a61-a739-5d6cd089ddad" (UID: "964cde77-c6b5-4a61-a739-5d6cd089ddad"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:10:54 crc kubenswrapper[4959]: I1007 14:10:54.321747 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/964cde77-c6b5-4a61-a739-5d6cd089ddad-kube-api-access-gc4f8" (OuterVolumeSpecName: "kube-api-access-gc4f8") pod "964cde77-c6b5-4a61-a739-5d6cd089ddad" (UID: "964cde77-c6b5-4a61-a739-5d6cd089ddad"). InnerVolumeSpecName "kube-api-access-gc4f8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:10:54 crc kubenswrapper[4959]: I1007 14:10:54.409898 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/964cde77-c6b5-4a61-a739-5d6cd089ddad-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "964cde77-c6b5-4a61-a739-5d6cd089ddad" (UID: "964cde77-c6b5-4a61-a739-5d6cd089ddad"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:10:54 crc kubenswrapper[4959]: I1007 14:10:54.415469 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gc4f8\" (UniqueName: \"kubernetes.io/projected/964cde77-c6b5-4a61-a739-5d6cd089ddad-kube-api-access-gc4f8\") on node \"crc\" DevicePath \"\"" Oct 07 14:10:54 crc kubenswrapper[4959]: I1007 14:10:54.415527 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/964cde77-c6b5-4a61-a739-5d6cd089ddad-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 14:10:54 crc kubenswrapper[4959]: I1007 14:10:54.415544 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/964cde77-c6b5-4a61-a739-5d6cd089ddad-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 14:10:54 crc kubenswrapper[4959]: I1007 14:10:54.745658 4959 generic.go:334] "Generic (PLEG): container finished" podID="964cde77-c6b5-4a61-a739-5d6cd089ddad" containerID="6e20f8686125a7697fbc5a68dcec1e5027ffa6b5d6a4b979fcd8245f94982da7" exitCode=0 Oct 07 14:10:54 crc kubenswrapper[4959]: I1007 14:10:54.745706 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7b2n7" event={"ID":"964cde77-c6b5-4a61-a739-5d6cd089ddad","Type":"ContainerDied","Data":"6e20f8686125a7697fbc5a68dcec1e5027ffa6b5d6a4b979fcd8245f94982da7"} Oct 07 14:10:54 crc kubenswrapper[4959]: I1007 14:10:54.746770 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7b2n7" event={"ID":"964cde77-c6b5-4a61-a739-5d6cd089ddad","Type":"ContainerDied","Data":"0506199fc3d409aa53fb96796b07ffe1860c33d19c30415df2914fe562ba555e"} Oct 07 14:10:54 crc kubenswrapper[4959]: I1007 14:10:54.745749 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7b2n7" Oct 07 14:10:54 crc kubenswrapper[4959]: I1007 14:10:54.746798 4959 scope.go:117] "RemoveContainer" containerID="6e20f8686125a7697fbc5a68dcec1e5027ffa6b5d6a4b979fcd8245f94982da7" Oct 07 14:10:54 crc kubenswrapper[4959]: I1007 14:10:54.777232 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7b2n7"] Oct 07 14:10:54 crc kubenswrapper[4959]: I1007 14:10:54.779450 4959 scope.go:117] "RemoveContainer" containerID="7895595e3c229042606a66ca2da681f0c45c7a9cce4918fb6da2d7ba37a01f5d" Oct 07 14:10:54 crc kubenswrapper[4959]: I1007 14:10:54.784740 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7b2n7"] Oct 07 14:10:54 crc kubenswrapper[4959]: I1007 14:10:54.809235 4959 scope.go:117] "RemoveContainer" containerID="baa35a208bf268fdc594266b821d2223bed497cc913d642a1504ce79b97c3a7b" Oct 07 14:10:54 crc kubenswrapper[4959]: I1007 14:10:54.863470 4959 scope.go:117] "RemoveContainer" containerID="6e20f8686125a7697fbc5a68dcec1e5027ffa6b5d6a4b979fcd8245f94982da7" Oct 07 14:10:54 crc kubenswrapper[4959]: E1007 14:10:54.863945 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e20f8686125a7697fbc5a68dcec1e5027ffa6b5d6a4b979fcd8245f94982da7\": container with ID starting with 6e20f8686125a7697fbc5a68dcec1e5027ffa6b5d6a4b979fcd8245f94982da7 not found: ID does not exist" containerID="6e20f8686125a7697fbc5a68dcec1e5027ffa6b5d6a4b979fcd8245f94982da7" Oct 07 14:10:54 crc kubenswrapper[4959]: I1007 14:10:54.863995 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e20f8686125a7697fbc5a68dcec1e5027ffa6b5d6a4b979fcd8245f94982da7"} err="failed to get container status \"6e20f8686125a7697fbc5a68dcec1e5027ffa6b5d6a4b979fcd8245f94982da7\": rpc error: code = NotFound desc = could not find container \"6e20f8686125a7697fbc5a68dcec1e5027ffa6b5d6a4b979fcd8245f94982da7\": container with ID starting with 6e20f8686125a7697fbc5a68dcec1e5027ffa6b5d6a4b979fcd8245f94982da7 not found: ID does not exist" Oct 07 14:10:54 crc kubenswrapper[4959]: I1007 14:10:54.864031 4959 scope.go:117] "RemoveContainer" containerID="7895595e3c229042606a66ca2da681f0c45c7a9cce4918fb6da2d7ba37a01f5d" Oct 07 14:10:54 crc kubenswrapper[4959]: E1007 14:10:54.864359 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7895595e3c229042606a66ca2da681f0c45c7a9cce4918fb6da2d7ba37a01f5d\": container with ID starting with 7895595e3c229042606a66ca2da681f0c45c7a9cce4918fb6da2d7ba37a01f5d not found: ID does not exist" containerID="7895595e3c229042606a66ca2da681f0c45c7a9cce4918fb6da2d7ba37a01f5d" Oct 07 14:10:54 crc kubenswrapper[4959]: I1007 14:10:54.864392 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7895595e3c229042606a66ca2da681f0c45c7a9cce4918fb6da2d7ba37a01f5d"} err="failed to get container status \"7895595e3c229042606a66ca2da681f0c45c7a9cce4918fb6da2d7ba37a01f5d\": rpc error: code = NotFound desc = could not find container \"7895595e3c229042606a66ca2da681f0c45c7a9cce4918fb6da2d7ba37a01f5d\": container with ID starting with 7895595e3c229042606a66ca2da681f0c45c7a9cce4918fb6da2d7ba37a01f5d not found: ID does not exist" Oct 07 14:10:54 crc kubenswrapper[4959]: I1007 14:10:54.864411 4959 scope.go:117] "RemoveContainer" containerID="baa35a208bf268fdc594266b821d2223bed497cc913d642a1504ce79b97c3a7b" Oct 07 14:10:54 crc kubenswrapper[4959]: E1007 14:10:54.864697 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"baa35a208bf268fdc594266b821d2223bed497cc913d642a1504ce79b97c3a7b\": container with ID starting with baa35a208bf268fdc594266b821d2223bed497cc913d642a1504ce79b97c3a7b not found: ID does not exist" containerID="baa35a208bf268fdc594266b821d2223bed497cc913d642a1504ce79b97c3a7b" Oct 07 14:10:54 crc kubenswrapper[4959]: I1007 14:10:54.864727 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"baa35a208bf268fdc594266b821d2223bed497cc913d642a1504ce79b97c3a7b"} err="failed to get container status \"baa35a208bf268fdc594266b821d2223bed497cc913d642a1504ce79b97c3a7b\": rpc error: code = NotFound desc = could not find container \"baa35a208bf268fdc594266b821d2223bed497cc913d642a1504ce79b97c3a7b\": container with ID starting with baa35a208bf268fdc594266b821d2223bed497cc913d642a1504ce79b97c3a7b not found: ID does not exist" Oct 07 14:10:56 crc kubenswrapper[4959]: I1007 14:10:56.667750 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="964cde77-c6b5-4a61-a739-5d6cd089ddad" path="/var/lib/kubelet/pods/964cde77-c6b5-4a61-a739-5d6cd089ddad/volumes" Oct 07 14:11:05 crc kubenswrapper[4959]: I1007 14:11:05.654170 4959 scope.go:117] "RemoveContainer" containerID="8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1" Oct 07 14:11:05 crc kubenswrapper[4959]: E1007 14:11:05.655039 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:11:13 crc kubenswrapper[4959]: I1007 14:11:13.923366 4959 generic.go:334] "Generic (PLEG): container finished" podID="e7967b3f-3a72-47d4-935d-4fa777b0fc46" containerID="35091a54cbb649c1df69456b4413f744e19f37ada482440213a5881c959b6079" exitCode=0 Oct 07 14:11:13 crc kubenswrapper[4959]: I1007 14:11:13.924183 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4" event={"ID":"e7967b3f-3a72-47d4-935d-4fa777b0fc46","Type":"ContainerDied","Data":"35091a54cbb649c1df69456b4413f744e19f37ada482440213a5881c959b6079"} Oct 07 14:11:15 crc kubenswrapper[4959]: I1007 14:11:15.325219 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4" Oct 07 14:11:15 crc kubenswrapper[4959]: I1007 14:11:15.405149 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e7967b3f-3a72-47d4-935d-4fa777b0fc46-ssh-key\") pod \"e7967b3f-3a72-47d4-935d-4fa777b0fc46\" (UID: \"e7967b3f-3a72-47d4-935d-4fa777b0fc46\") " Oct 07 14:11:15 crc kubenswrapper[4959]: I1007 14:11:15.405343 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxwcb\" (UniqueName: \"kubernetes.io/projected/e7967b3f-3a72-47d4-935d-4fa777b0fc46-kube-api-access-mxwcb\") pod \"e7967b3f-3a72-47d4-935d-4fa777b0fc46\" (UID: \"e7967b3f-3a72-47d4-935d-4fa777b0fc46\") " Oct 07 14:11:15 crc kubenswrapper[4959]: I1007 14:11:15.405364 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7967b3f-3a72-47d4-935d-4fa777b0fc46-bootstrap-combined-ca-bundle\") pod \"e7967b3f-3a72-47d4-935d-4fa777b0fc46\" (UID: \"e7967b3f-3a72-47d4-935d-4fa777b0fc46\") " Oct 07 14:11:15 crc kubenswrapper[4959]: I1007 14:11:15.405393 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e7967b3f-3a72-47d4-935d-4fa777b0fc46-inventory\") pod \"e7967b3f-3a72-47d4-935d-4fa777b0fc46\" (UID: \"e7967b3f-3a72-47d4-935d-4fa777b0fc46\") " Oct 07 14:11:15 crc kubenswrapper[4959]: I1007 14:11:15.419367 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7967b3f-3a72-47d4-935d-4fa777b0fc46-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "e7967b3f-3a72-47d4-935d-4fa777b0fc46" (UID: "e7967b3f-3a72-47d4-935d-4fa777b0fc46"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:11:15 crc kubenswrapper[4959]: I1007 14:11:15.424985 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7967b3f-3a72-47d4-935d-4fa777b0fc46-kube-api-access-mxwcb" (OuterVolumeSpecName: "kube-api-access-mxwcb") pod "e7967b3f-3a72-47d4-935d-4fa777b0fc46" (UID: "e7967b3f-3a72-47d4-935d-4fa777b0fc46"). InnerVolumeSpecName "kube-api-access-mxwcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:11:15 crc kubenswrapper[4959]: I1007 14:11:15.438897 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7967b3f-3a72-47d4-935d-4fa777b0fc46-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e7967b3f-3a72-47d4-935d-4fa777b0fc46" (UID: "e7967b3f-3a72-47d4-935d-4fa777b0fc46"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:11:15 crc kubenswrapper[4959]: I1007 14:11:15.442454 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7967b3f-3a72-47d4-935d-4fa777b0fc46-inventory" (OuterVolumeSpecName: "inventory") pod "e7967b3f-3a72-47d4-935d-4fa777b0fc46" (UID: "e7967b3f-3a72-47d4-935d-4fa777b0fc46"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:11:15 crc kubenswrapper[4959]: I1007 14:11:15.507974 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e7967b3f-3a72-47d4-935d-4fa777b0fc46-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 14:11:15 crc kubenswrapper[4959]: I1007 14:11:15.508012 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e7967b3f-3a72-47d4-935d-4fa777b0fc46-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 14:11:15 crc kubenswrapper[4959]: I1007 14:11:15.508024 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxwcb\" (UniqueName: \"kubernetes.io/projected/e7967b3f-3a72-47d4-935d-4fa777b0fc46-kube-api-access-mxwcb\") on node \"crc\" DevicePath \"\"" Oct 07 14:11:15 crc kubenswrapper[4959]: I1007 14:11:15.508039 4959 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7967b3f-3a72-47d4-935d-4fa777b0fc46-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:11:15 crc kubenswrapper[4959]: I1007 14:11:15.951130 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4" event={"ID":"e7967b3f-3a72-47d4-935d-4fa777b0fc46","Type":"ContainerDied","Data":"617cc0853cf664e5b5efc77aef87b5e40f157abc4341e6a9be07ee1b25cb47a8"} Oct 07 14:11:15 crc kubenswrapper[4959]: I1007 14:11:15.951722 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="617cc0853cf664e5b5efc77aef87b5e40f157abc4341e6a9be07ee1b25cb47a8" Oct 07 14:11:15 crc kubenswrapper[4959]: I1007 14:11:15.951287 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4" Oct 07 14:11:16 crc kubenswrapper[4959]: I1007 14:11:16.030010 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q"] Oct 07 14:11:16 crc kubenswrapper[4959]: E1007 14:11:16.030375 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7967b3f-3a72-47d4-935d-4fa777b0fc46" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 07 14:11:16 crc kubenswrapper[4959]: I1007 14:11:16.030395 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7967b3f-3a72-47d4-935d-4fa777b0fc46" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 07 14:11:16 crc kubenswrapper[4959]: E1007 14:11:16.030415 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="964cde77-c6b5-4a61-a739-5d6cd089ddad" containerName="extract-utilities" Oct 07 14:11:16 crc kubenswrapper[4959]: I1007 14:11:16.030423 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="964cde77-c6b5-4a61-a739-5d6cd089ddad" containerName="extract-utilities" Oct 07 14:11:16 crc kubenswrapper[4959]: E1007 14:11:16.030453 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="964cde77-c6b5-4a61-a739-5d6cd089ddad" containerName="registry-server" Oct 07 14:11:16 crc kubenswrapper[4959]: I1007 14:11:16.030459 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="964cde77-c6b5-4a61-a739-5d6cd089ddad" containerName="registry-server" Oct 07 14:11:16 crc kubenswrapper[4959]: E1007 14:11:16.030470 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="964cde77-c6b5-4a61-a739-5d6cd089ddad" containerName="extract-content" Oct 07 14:11:16 crc kubenswrapper[4959]: I1007 14:11:16.030477 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="964cde77-c6b5-4a61-a739-5d6cd089ddad" containerName="extract-content" Oct 07 14:11:16 crc kubenswrapper[4959]: I1007 14:11:16.030626 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7967b3f-3a72-47d4-935d-4fa777b0fc46" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 07 14:11:16 crc kubenswrapper[4959]: I1007 14:11:16.030641 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="964cde77-c6b5-4a61-a739-5d6cd089ddad" containerName="registry-server" Oct 07 14:11:16 crc kubenswrapper[4959]: I1007 14:11:16.031319 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q" Oct 07 14:11:16 crc kubenswrapper[4959]: I1007 14:11:16.035582 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 14:11:16 crc kubenswrapper[4959]: I1007 14:11:16.035773 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 14:11:16 crc kubenswrapper[4959]: I1007 14:11:16.035842 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 14:11:16 crc kubenswrapper[4959]: I1007 14:11:16.035931 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-c6vmh" Oct 07 14:11:16 crc kubenswrapper[4959]: I1007 14:11:16.044637 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q"] Oct 07 14:11:16 crc kubenswrapper[4959]: I1007 14:11:16.119434 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d843db5d-75e6-4f95-a297-41af3451648f-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q\" (UID: \"d843db5d-75e6-4f95-a297-41af3451648f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q" Oct 07 14:11:16 crc kubenswrapper[4959]: I1007 14:11:16.119773 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6f27m\" (UniqueName: \"kubernetes.io/projected/d843db5d-75e6-4f95-a297-41af3451648f-kube-api-access-6f27m\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q\" (UID: \"d843db5d-75e6-4f95-a297-41af3451648f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q" Oct 07 14:11:16 crc kubenswrapper[4959]: I1007 14:11:16.120016 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d843db5d-75e6-4f95-a297-41af3451648f-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q\" (UID: \"d843db5d-75e6-4f95-a297-41af3451648f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q" Oct 07 14:11:16 crc kubenswrapper[4959]: I1007 14:11:16.222200 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6f27m\" (UniqueName: \"kubernetes.io/projected/d843db5d-75e6-4f95-a297-41af3451648f-kube-api-access-6f27m\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q\" (UID: \"d843db5d-75e6-4f95-a297-41af3451648f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q" Oct 07 14:11:16 crc kubenswrapper[4959]: I1007 14:11:16.222290 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d843db5d-75e6-4f95-a297-41af3451648f-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q\" (UID: \"d843db5d-75e6-4f95-a297-41af3451648f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q" Oct 07 14:11:16 crc kubenswrapper[4959]: I1007 14:11:16.222428 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d843db5d-75e6-4f95-a297-41af3451648f-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q\" (UID: \"d843db5d-75e6-4f95-a297-41af3451648f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q" Oct 07 14:11:16 crc kubenswrapper[4959]: I1007 14:11:16.226988 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d843db5d-75e6-4f95-a297-41af3451648f-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q\" (UID: \"d843db5d-75e6-4f95-a297-41af3451648f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q" Oct 07 14:11:16 crc kubenswrapper[4959]: I1007 14:11:16.229744 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d843db5d-75e6-4f95-a297-41af3451648f-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q\" (UID: \"d843db5d-75e6-4f95-a297-41af3451648f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q" Oct 07 14:11:16 crc kubenswrapper[4959]: I1007 14:11:16.242878 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6f27m\" (UniqueName: \"kubernetes.io/projected/d843db5d-75e6-4f95-a297-41af3451648f-kube-api-access-6f27m\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q\" (UID: \"d843db5d-75e6-4f95-a297-41af3451648f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q" Oct 07 14:11:16 crc kubenswrapper[4959]: I1007 14:11:16.349745 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q" Oct 07 14:11:16 crc kubenswrapper[4959]: I1007 14:11:16.886431 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q"] Oct 07 14:11:16 crc kubenswrapper[4959]: I1007 14:11:16.895746 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 14:11:16 crc kubenswrapper[4959]: I1007 14:11:16.960937 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q" event={"ID":"d843db5d-75e6-4f95-a297-41af3451648f","Type":"ContainerStarted","Data":"f2fa062db5a3a7e3bb28c9874850decf41d5affa6611dfa06d9eef3e02c21f1e"} Oct 07 14:11:18 crc kubenswrapper[4959]: I1007 14:11:18.666811 4959 scope.go:117] "RemoveContainer" containerID="8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1" Oct 07 14:11:18 crc kubenswrapper[4959]: E1007 14:11:18.668457 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:11:18 crc kubenswrapper[4959]: I1007 14:11:18.985888 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q" event={"ID":"d843db5d-75e6-4f95-a297-41af3451648f","Type":"ContainerStarted","Data":"1c5f573eb9c4a3ab7894dbab43f9e66531c2be377ebf98a67796c9c94b35f235"} Oct 07 14:11:19 crc kubenswrapper[4959]: I1007 14:11:19.006818 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q" podStartSLOduration=2.168342356 podStartE2EDuration="3.006798252s" podCreationTimestamp="2025-10-07 14:11:16 +0000 UTC" firstStartedPulling="2025-10-07 14:11:16.89531117 +0000 UTC m=+1558.978715495" lastFinishedPulling="2025-10-07 14:11:17.733767066 +0000 UTC m=+1559.817171391" observedRunningTime="2025-10-07 14:11:19.000203828 +0000 UTC m=+1561.083608173" watchObservedRunningTime="2025-10-07 14:11:19.006798252 +0000 UTC m=+1561.090202577" Oct 07 14:11:29 crc kubenswrapper[4959]: I1007 14:11:29.653256 4959 scope.go:117] "RemoveContainer" containerID="8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1" Oct 07 14:11:29 crc kubenswrapper[4959]: E1007 14:11:29.653824 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:11:40 crc kubenswrapper[4959]: I1007 14:11:40.653852 4959 scope.go:117] "RemoveContainer" containerID="8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1" Oct 07 14:11:40 crc kubenswrapper[4959]: E1007 14:11:40.654603 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:11:47 crc kubenswrapper[4959]: I1007 14:11:47.033807 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-jgf7k"] Oct 07 14:11:47 crc kubenswrapper[4959]: I1007 14:11:47.041497 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-jgf7k"] Oct 07 14:11:48 crc kubenswrapper[4959]: I1007 14:11:48.053013 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-4k9wm"] Oct 07 14:11:48 crc kubenswrapper[4959]: I1007 14:11:48.065026 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-4k9wm"] Oct 07 14:11:48 crc kubenswrapper[4959]: I1007 14:11:48.070607 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-sxjr4"] Oct 07 14:11:48 crc kubenswrapper[4959]: I1007 14:11:48.078453 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-sxjr4"] Oct 07 14:11:48 crc kubenswrapper[4959]: I1007 14:11:48.668493 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67166a6a-e8f0-479e-acda-1fec5161eac4" path="/var/lib/kubelet/pods/67166a6a-e8f0-479e-acda-1fec5161eac4/volumes" Oct 07 14:11:48 crc kubenswrapper[4959]: I1007 14:11:48.669032 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2ed38ec-6665-4e4f-9a10-1eb8c958a306" path="/var/lib/kubelet/pods/b2ed38ec-6665-4e4f-9a10-1eb8c958a306/volumes" Oct 07 14:11:48 crc kubenswrapper[4959]: I1007 14:11:48.669505 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbe13445-5cad-4de1-aed4-6c95c41c83b2" path="/var/lib/kubelet/pods/bbe13445-5cad-4de1-aed4-6c95c41c83b2/volumes" Oct 07 14:11:51 crc kubenswrapper[4959]: I1007 14:11:51.653524 4959 scope.go:117] "RemoveContainer" containerID="8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1" Oct 07 14:11:51 crc kubenswrapper[4959]: E1007 14:11:51.654128 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:11:57 crc kubenswrapper[4959]: I1007 14:11:57.027383 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-6996-account-create-pvzth"] Oct 07 14:11:57 crc kubenswrapper[4959]: I1007 14:11:57.036423 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-6996-account-create-pvzth"] Oct 07 14:11:58 crc kubenswrapper[4959]: I1007 14:11:58.673978 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8161816-d8b5-4e8a-a132-3ae55dcff880" path="/var/lib/kubelet/pods/e8161816-d8b5-4e8a-a132-3ae55dcff880/volumes" Oct 07 14:12:02 crc kubenswrapper[4959]: I1007 14:12:02.653602 4959 scope.go:117] "RemoveContainer" containerID="8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1" Oct 07 14:12:02 crc kubenswrapper[4959]: E1007 14:12:02.654404 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:12:04 crc kubenswrapper[4959]: I1007 14:12:04.125840 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gtfgj"] Oct 07 14:12:04 crc kubenswrapper[4959]: I1007 14:12:04.128570 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gtfgj" Oct 07 14:12:04 crc kubenswrapper[4959]: I1007 14:12:04.177504 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gtfgj"] Oct 07 14:12:04 crc kubenswrapper[4959]: I1007 14:12:04.308258 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63dd100a-8db8-4adb-810c-dc2c06a7371a-utilities\") pod \"community-operators-gtfgj\" (UID: \"63dd100a-8db8-4adb-810c-dc2c06a7371a\") " pod="openshift-marketplace/community-operators-gtfgj" Oct 07 14:12:04 crc kubenswrapper[4959]: I1007 14:12:04.308387 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63dd100a-8db8-4adb-810c-dc2c06a7371a-catalog-content\") pod \"community-operators-gtfgj\" (UID: \"63dd100a-8db8-4adb-810c-dc2c06a7371a\") " pod="openshift-marketplace/community-operators-gtfgj" Oct 07 14:12:04 crc kubenswrapper[4959]: I1007 14:12:04.308468 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjl6g\" (UniqueName: \"kubernetes.io/projected/63dd100a-8db8-4adb-810c-dc2c06a7371a-kube-api-access-tjl6g\") pod \"community-operators-gtfgj\" (UID: \"63dd100a-8db8-4adb-810c-dc2c06a7371a\") " pod="openshift-marketplace/community-operators-gtfgj" Oct 07 14:12:04 crc kubenswrapper[4959]: I1007 14:12:04.409717 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjl6g\" (UniqueName: \"kubernetes.io/projected/63dd100a-8db8-4adb-810c-dc2c06a7371a-kube-api-access-tjl6g\") pod \"community-operators-gtfgj\" (UID: \"63dd100a-8db8-4adb-810c-dc2c06a7371a\") " pod="openshift-marketplace/community-operators-gtfgj" Oct 07 14:12:04 crc kubenswrapper[4959]: I1007 14:12:04.409815 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63dd100a-8db8-4adb-810c-dc2c06a7371a-utilities\") pod \"community-operators-gtfgj\" (UID: \"63dd100a-8db8-4adb-810c-dc2c06a7371a\") " pod="openshift-marketplace/community-operators-gtfgj" Oct 07 14:12:04 crc kubenswrapper[4959]: I1007 14:12:04.409872 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63dd100a-8db8-4adb-810c-dc2c06a7371a-catalog-content\") pod \"community-operators-gtfgj\" (UID: \"63dd100a-8db8-4adb-810c-dc2c06a7371a\") " pod="openshift-marketplace/community-operators-gtfgj" Oct 07 14:12:04 crc kubenswrapper[4959]: I1007 14:12:04.410383 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63dd100a-8db8-4adb-810c-dc2c06a7371a-catalog-content\") pod \"community-operators-gtfgj\" (UID: \"63dd100a-8db8-4adb-810c-dc2c06a7371a\") " pod="openshift-marketplace/community-operators-gtfgj" Oct 07 14:12:04 crc kubenswrapper[4959]: I1007 14:12:04.410484 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63dd100a-8db8-4adb-810c-dc2c06a7371a-utilities\") pod \"community-operators-gtfgj\" (UID: \"63dd100a-8db8-4adb-810c-dc2c06a7371a\") " pod="openshift-marketplace/community-operators-gtfgj" Oct 07 14:12:04 crc kubenswrapper[4959]: I1007 14:12:04.438023 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjl6g\" (UniqueName: \"kubernetes.io/projected/63dd100a-8db8-4adb-810c-dc2c06a7371a-kube-api-access-tjl6g\") pod \"community-operators-gtfgj\" (UID: \"63dd100a-8db8-4adb-810c-dc2c06a7371a\") " pod="openshift-marketplace/community-operators-gtfgj" Oct 07 14:12:04 crc kubenswrapper[4959]: I1007 14:12:04.493122 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gtfgj" Oct 07 14:12:04 crc kubenswrapper[4959]: I1007 14:12:04.975585 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gtfgj"] Oct 07 14:12:05 crc kubenswrapper[4959]: I1007 14:12:05.031189 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-c1a2-account-create-xbdrw"] Oct 07 14:12:05 crc kubenswrapper[4959]: I1007 14:12:05.041069 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-c1a2-account-create-xbdrw"] Oct 07 14:12:05 crc kubenswrapper[4959]: I1007 14:12:05.450724 4959 generic.go:334] "Generic (PLEG): container finished" podID="63dd100a-8db8-4adb-810c-dc2c06a7371a" containerID="b85c6e5fc7126e5742cb1ef996905d4682e00c23db9f48d33d98b9e17b722c07" exitCode=0 Oct 07 14:12:05 crc kubenswrapper[4959]: I1007 14:12:05.450775 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gtfgj" event={"ID":"63dd100a-8db8-4adb-810c-dc2c06a7371a","Type":"ContainerDied","Data":"b85c6e5fc7126e5742cb1ef996905d4682e00c23db9f48d33d98b9e17b722c07"} Oct 07 14:12:05 crc kubenswrapper[4959]: I1007 14:12:05.451035 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gtfgj" event={"ID":"63dd100a-8db8-4adb-810c-dc2c06a7371a","Type":"ContainerStarted","Data":"f7298abf0a370d3528785e6d5c1e91bbbd1d3178d6645b7769df2c02f5938650"} Oct 07 14:12:06 crc kubenswrapper[4959]: I1007 14:12:06.030667 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-43c9-account-create-m7dhb"] Oct 07 14:12:06 crc kubenswrapper[4959]: I1007 14:12:06.037755 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-43c9-account-create-m7dhb"] Oct 07 14:12:06 crc kubenswrapper[4959]: I1007 14:12:06.685722 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a2d8750-d82d-4a23-ac9f-20e4331381e3" path="/var/lib/kubelet/pods/6a2d8750-d82d-4a23-ac9f-20e4331381e3/volumes" Oct 07 14:12:06 crc kubenswrapper[4959]: I1007 14:12:06.686663 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccf90649-6706-47f5-a899-f8674a14efcd" path="/var/lib/kubelet/pods/ccf90649-6706-47f5-a899-f8674a14efcd/volumes" Oct 07 14:12:07 crc kubenswrapper[4959]: I1007 14:12:07.472018 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gtfgj" event={"ID":"63dd100a-8db8-4adb-810c-dc2c06a7371a","Type":"ContainerStarted","Data":"83cb54745aed5f37f6bbfa6bfb623c3b14335e1e762384c3ecea81c8b47d3011"} Oct 07 14:12:08 crc kubenswrapper[4959]: I1007 14:12:08.483453 4959 generic.go:334] "Generic (PLEG): container finished" podID="63dd100a-8db8-4adb-810c-dc2c06a7371a" containerID="83cb54745aed5f37f6bbfa6bfb623c3b14335e1e762384c3ecea81c8b47d3011" exitCode=0 Oct 07 14:12:08 crc kubenswrapper[4959]: I1007 14:12:08.483516 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gtfgj" event={"ID":"63dd100a-8db8-4adb-810c-dc2c06a7371a","Type":"ContainerDied","Data":"83cb54745aed5f37f6bbfa6bfb623c3b14335e1e762384c3ecea81c8b47d3011"} Oct 07 14:12:09 crc kubenswrapper[4959]: I1007 14:12:09.492125 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gtfgj" event={"ID":"63dd100a-8db8-4adb-810c-dc2c06a7371a","Type":"ContainerStarted","Data":"621342aa753ca1e6345cf745fc512ba9559e065d5ce1015ac0cfca4910abb9ea"} Oct 07 14:12:09 crc kubenswrapper[4959]: I1007 14:12:09.516400 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gtfgj" podStartSLOduration=2.021738273 podStartE2EDuration="5.516382907s" podCreationTimestamp="2025-10-07 14:12:04 +0000 UTC" firstStartedPulling="2025-10-07 14:12:05.452122167 +0000 UTC m=+1607.535526492" lastFinishedPulling="2025-10-07 14:12:08.946766791 +0000 UTC m=+1611.030171126" observedRunningTime="2025-10-07 14:12:09.50688245 +0000 UTC m=+1611.590286775" watchObservedRunningTime="2025-10-07 14:12:09.516382907 +0000 UTC m=+1611.599787232" Oct 07 14:12:14 crc kubenswrapper[4959]: I1007 14:12:14.493634 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gtfgj" Oct 07 14:12:14 crc kubenswrapper[4959]: I1007 14:12:14.494266 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gtfgj" Oct 07 14:12:14 crc kubenswrapper[4959]: I1007 14:12:14.547852 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gtfgj" Oct 07 14:12:14 crc kubenswrapper[4959]: I1007 14:12:14.593526 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gtfgj" Oct 07 14:12:14 crc kubenswrapper[4959]: I1007 14:12:14.782657 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gtfgj"] Oct 07 14:12:16 crc kubenswrapper[4959]: I1007 14:12:16.555843 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gtfgj" podUID="63dd100a-8db8-4adb-810c-dc2c06a7371a" containerName="registry-server" containerID="cri-o://621342aa753ca1e6345cf745fc512ba9559e065d5ce1015ac0cfca4910abb9ea" gracePeriod=2 Oct 07 14:12:16 crc kubenswrapper[4959]: I1007 14:12:16.653306 4959 scope.go:117] "RemoveContainer" containerID="8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1" Oct 07 14:12:16 crc kubenswrapper[4959]: E1007 14:12:16.653553 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:12:17 crc kubenswrapper[4959]: I1007 14:12:17.032383 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gtfgj" Oct 07 14:12:17 crc kubenswrapper[4959]: I1007 14:12:17.142549 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63dd100a-8db8-4adb-810c-dc2c06a7371a-utilities\") pod \"63dd100a-8db8-4adb-810c-dc2c06a7371a\" (UID: \"63dd100a-8db8-4adb-810c-dc2c06a7371a\") " Oct 07 14:12:17 crc kubenswrapper[4959]: I1007 14:12:17.142639 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63dd100a-8db8-4adb-810c-dc2c06a7371a-catalog-content\") pod \"63dd100a-8db8-4adb-810c-dc2c06a7371a\" (UID: \"63dd100a-8db8-4adb-810c-dc2c06a7371a\") " Oct 07 14:12:17 crc kubenswrapper[4959]: I1007 14:12:17.142762 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjl6g\" (UniqueName: \"kubernetes.io/projected/63dd100a-8db8-4adb-810c-dc2c06a7371a-kube-api-access-tjl6g\") pod \"63dd100a-8db8-4adb-810c-dc2c06a7371a\" (UID: \"63dd100a-8db8-4adb-810c-dc2c06a7371a\") " Oct 07 14:12:17 crc kubenswrapper[4959]: I1007 14:12:17.145208 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63dd100a-8db8-4adb-810c-dc2c06a7371a-utilities" (OuterVolumeSpecName: "utilities") pod "63dd100a-8db8-4adb-810c-dc2c06a7371a" (UID: "63dd100a-8db8-4adb-810c-dc2c06a7371a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:12:17 crc kubenswrapper[4959]: I1007 14:12:17.152598 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63dd100a-8db8-4adb-810c-dc2c06a7371a-kube-api-access-tjl6g" (OuterVolumeSpecName: "kube-api-access-tjl6g") pod "63dd100a-8db8-4adb-810c-dc2c06a7371a" (UID: "63dd100a-8db8-4adb-810c-dc2c06a7371a"). InnerVolumeSpecName "kube-api-access-tjl6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:12:17 crc kubenswrapper[4959]: I1007 14:12:17.203469 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63dd100a-8db8-4adb-810c-dc2c06a7371a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "63dd100a-8db8-4adb-810c-dc2c06a7371a" (UID: "63dd100a-8db8-4adb-810c-dc2c06a7371a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:12:17 crc kubenswrapper[4959]: I1007 14:12:17.245256 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63dd100a-8db8-4adb-810c-dc2c06a7371a-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 14:12:17 crc kubenswrapper[4959]: I1007 14:12:17.245301 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63dd100a-8db8-4adb-810c-dc2c06a7371a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 14:12:17 crc kubenswrapper[4959]: I1007 14:12:17.245318 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjl6g\" (UniqueName: \"kubernetes.io/projected/63dd100a-8db8-4adb-810c-dc2c06a7371a-kube-api-access-tjl6g\") on node \"crc\" DevicePath \"\"" Oct 07 14:12:17 crc kubenswrapper[4959]: I1007 14:12:17.567231 4959 generic.go:334] "Generic (PLEG): container finished" podID="63dd100a-8db8-4adb-810c-dc2c06a7371a" containerID="621342aa753ca1e6345cf745fc512ba9559e065d5ce1015ac0cfca4910abb9ea" exitCode=0 Oct 07 14:12:17 crc kubenswrapper[4959]: I1007 14:12:17.567294 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gtfgj" Oct 07 14:12:17 crc kubenswrapper[4959]: I1007 14:12:17.567305 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gtfgj" event={"ID":"63dd100a-8db8-4adb-810c-dc2c06a7371a","Type":"ContainerDied","Data":"621342aa753ca1e6345cf745fc512ba9559e065d5ce1015ac0cfca4910abb9ea"} Oct 07 14:12:17 crc kubenswrapper[4959]: I1007 14:12:17.567360 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gtfgj" event={"ID":"63dd100a-8db8-4adb-810c-dc2c06a7371a","Type":"ContainerDied","Data":"f7298abf0a370d3528785e6d5c1e91bbbd1d3178d6645b7769df2c02f5938650"} Oct 07 14:12:17 crc kubenswrapper[4959]: I1007 14:12:17.567397 4959 scope.go:117] "RemoveContainer" containerID="621342aa753ca1e6345cf745fc512ba9559e065d5ce1015ac0cfca4910abb9ea" Oct 07 14:12:17 crc kubenswrapper[4959]: I1007 14:12:17.588965 4959 scope.go:117] "RemoveContainer" containerID="83cb54745aed5f37f6bbfa6bfb623c3b14335e1e762384c3ecea81c8b47d3011" Oct 07 14:12:17 crc kubenswrapper[4959]: I1007 14:12:17.607133 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gtfgj"] Oct 07 14:12:17 crc kubenswrapper[4959]: I1007 14:12:17.617157 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gtfgj"] Oct 07 14:12:17 crc kubenswrapper[4959]: I1007 14:12:17.624648 4959 scope.go:117] "RemoveContainer" containerID="b85c6e5fc7126e5742cb1ef996905d4682e00c23db9f48d33d98b9e17b722c07" Oct 07 14:12:17 crc kubenswrapper[4959]: I1007 14:12:17.663916 4959 scope.go:117] "RemoveContainer" containerID="621342aa753ca1e6345cf745fc512ba9559e065d5ce1015ac0cfca4910abb9ea" Oct 07 14:12:17 crc kubenswrapper[4959]: E1007 14:12:17.664413 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"621342aa753ca1e6345cf745fc512ba9559e065d5ce1015ac0cfca4910abb9ea\": container with ID starting with 621342aa753ca1e6345cf745fc512ba9559e065d5ce1015ac0cfca4910abb9ea not found: ID does not exist" containerID="621342aa753ca1e6345cf745fc512ba9559e065d5ce1015ac0cfca4910abb9ea" Oct 07 14:12:17 crc kubenswrapper[4959]: I1007 14:12:17.664444 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"621342aa753ca1e6345cf745fc512ba9559e065d5ce1015ac0cfca4910abb9ea"} err="failed to get container status \"621342aa753ca1e6345cf745fc512ba9559e065d5ce1015ac0cfca4910abb9ea\": rpc error: code = NotFound desc = could not find container \"621342aa753ca1e6345cf745fc512ba9559e065d5ce1015ac0cfca4910abb9ea\": container with ID starting with 621342aa753ca1e6345cf745fc512ba9559e065d5ce1015ac0cfca4910abb9ea not found: ID does not exist" Oct 07 14:12:17 crc kubenswrapper[4959]: I1007 14:12:17.664488 4959 scope.go:117] "RemoveContainer" containerID="83cb54745aed5f37f6bbfa6bfb623c3b14335e1e762384c3ecea81c8b47d3011" Oct 07 14:12:17 crc kubenswrapper[4959]: E1007 14:12:17.665088 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83cb54745aed5f37f6bbfa6bfb623c3b14335e1e762384c3ecea81c8b47d3011\": container with ID starting with 83cb54745aed5f37f6bbfa6bfb623c3b14335e1e762384c3ecea81c8b47d3011 not found: ID does not exist" containerID="83cb54745aed5f37f6bbfa6bfb623c3b14335e1e762384c3ecea81c8b47d3011" Oct 07 14:12:17 crc kubenswrapper[4959]: I1007 14:12:17.665136 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83cb54745aed5f37f6bbfa6bfb623c3b14335e1e762384c3ecea81c8b47d3011"} err="failed to get container status \"83cb54745aed5f37f6bbfa6bfb623c3b14335e1e762384c3ecea81c8b47d3011\": rpc error: code = NotFound desc = could not find container \"83cb54745aed5f37f6bbfa6bfb623c3b14335e1e762384c3ecea81c8b47d3011\": container with ID starting with 83cb54745aed5f37f6bbfa6bfb623c3b14335e1e762384c3ecea81c8b47d3011 not found: ID does not exist" Oct 07 14:12:17 crc kubenswrapper[4959]: I1007 14:12:17.665156 4959 scope.go:117] "RemoveContainer" containerID="b85c6e5fc7126e5742cb1ef996905d4682e00c23db9f48d33d98b9e17b722c07" Oct 07 14:12:17 crc kubenswrapper[4959]: E1007 14:12:17.665421 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b85c6e5fc7126e5742cb1ef996905d4682e00c23db9f48d33d98b9e17b722c07\": container with ID starting with b85c6e5fc7126e5742cb1ef996905d4682e00c23db9f48d33d98b9e17b722c07 not found: ID does not exist" containerID="b85c6e5fc7126e5742cb1ef996905d4682e00c23db9f48d33d98b9e17b722c07" Oct 07 14:12:17 crc kubenswrapper[4959]: I1007 14:12:17.665522 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b85c6e5fc7126e5742cb1ef996905d4682e00c23db9f48d33d98b9e17b722c07"} err="failed to get container status \"b85c6e5fc7126e5742cb1ef996905d4682e00c23db9f48d33d98b9e17b722c07\": rpc error: code = NotFound desc = could not find container \"b85c6e5fc7126e5742cb1ef996905d4682e00c23db9f48d33d98b9e17b722c07\": container with ID starting with b85c6e5fc7126e5742cb1ef996905d4682e00c23db9f48d33d98b9e17b722c07 not found: ID does not exist" Oct 07 14:12:18 crc kubenswrapper[4959]: I1007 14:12:18.670063 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63dd100a-8db8-4adb-810c-dc2c06a7371a" path="/var/lib/kubelet/pods/63dd100a-8db8-4adb-810c-dc2c06a7371a/volumes" Oct 07 14:12:21 crc kubenswrapper[4959]: I1007 14:12:21.045012 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-bpqp9"] Oct 07 14:12:21 crc kubenswrapper[4959]: I1007 14:12:21.053058 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-btxq4"] Oct 07 14:12:21 crc kubenswrapper[4959]: I1007 14:12:21.064769 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-qgn9g"] Oct 07 14:12:21 crc kubenswrapper[4959]: I1007 14:12:21.072884 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-btxq4"] Oct 07 14:12:21 crc kubenswrapper[4959]: I1007 14:12:21.079935 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-qgn9g"] Oct 07 14:12:21 crc kubenswrapper[4959]: I1007 14:12:21.086316 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-bpqp9"] Oct 07 14:12:22 crc kubenswrapper[4959]: I1007 14:12:22.673922 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="010fee6b-bceb-4a25-b50f-19ccc1eb79a7" path="/var/lib/kubelet/pods/010fee6b-bceb-4a25-b50f-19ccc1eb79a7/volumes" Oct 07 14:12:22 crc kubenswrapper[4959]: I1007 14:12:22.674464 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bea0796-b66c-4c44-be6e-840bde657893" path="/var/lib/kubelet/pods/2bea0796-b66c-4c44-be6e-840bde657893/volumes" Oct 07 14:12:22 crc kubenswrapper[4959]: I1007 14:12:22.674929 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32df9b0d-1d2c-4724-b6c7-c4902baf8ae0" path="/var/lib/kubelet/pods/32df9b0d-1d2c-4724-b6c7-c4902baf8ae0/volumes" Oct 07 14:12:26 crc kubenswrapper[4959]: I1007 14:12:26.663461 4959 generic.go:334] "Generic (PLEG): container finished" podID="d843db5d-75e6-4f95-a297-41af3451648f" containerID="1c5f573eb9c4a3ab7894dbab43f9e66531c2be377ebf98a67796c9c94b35f235" exitCode=0 Oct 07 14:12:26 crc kubenswrapper[4959]: I1007 14:12:26.666423 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q" event={"ID":"d843db5d-75e6-4f95-a297-41af3451648f","Type":"ContainerDied","Data":"1c5f573eb9c4a3ab7894dbab43f9e66531c2be377ebf98a67796c9c94b35f235"} Oct 07 14:12:27 crc kubenswrapper[4959]: I1007 14:12:27.758070 4959 scope.go:117] "RemoveContainer" containerID="0138d038f76b86146c8dbb8e948f7c884f6720f31483a9b2900a96a1ec074d4a" Oct 07 14:12:27 crc kubenswrapper[4959]: I1007 14:12:27.781990 4959 scope.go:117] "RemoveContainer" containerID="852d37b59bb0d41d1e34f8f3b61f053265e9f6cd4f9ae4a11c4f2fd232721040" Oct 07 14:12:27 crc kubenswrapper[4959]: I1007 14:12:27.853424 4959 scope.go:117] "RemoveContainer" containerID="ee3fd2485575e455c60f1fb347f4e962b59a424640d5ff51f426bfe326a4546b" Oct 07 14:12:27 crc kubenswrapper[4959]: I1007 14:12:27.877745 4959 scope.go:117] "RemoveContainer" containerID="da70dc6d97f7d723f1e2cd16678e421ed5399ebcb0fb5cce68c6f17e94915a1e" Oct 07 14:12:27 crc kubenswrapper[4959]: I1007 14:12:27.937358 4959 scope.go:117] "RemoveContainer" containerID="16de1daa04529c2509767c249c3c96e41a4928eab63f324fd5b3fd33f8c4d01b" Oct 07 14:12:27 crc kubenswrapper[4959]: I1007 14:12:27.970879 4959 scope.go:117] "RemoveContainer" containerID="f652c5f885e6d1fdb8237b16d9ad7f2ecfd9fd09d3af2513f73f898bf90cadff" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.009483 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.015824 4959 scope.go:117] "RemoveContainer" containerID="1d7299eb35bfd1ec5e01a5d11f283b0e2b84c38ad8c0ca17eb953750bc8426a5" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.054403 4959 scope.go:117] "RemoveContainer" containerID="9fb6ccbbc199190c97616fe5220ff9a32382a1179e278b4f7caa3d4e2e13f6ec" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.068472 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d843db5d-75e6-4f95-a297-41af3451648f-inventory\") pod \"d843db5d-75e6-4f95-a297-41af3451648f\" (UID: \"d843db5d-75e6-4f95-a297-41af3451648f\") " Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.068972 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d843db5d-75e6-4f95-a297-41af3451648f-ssh-key\") pod \"d843db5d-75e6-4f95-a297-41af3451648f\" (UID: \"d843db5d-75e6-4f95-a297-41af3451648f\") " Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.069064 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6f27m\" (UniqueName: \"kubernetes.io/projected/d843db5d-75e6-4f95-a297-41af3451648f-kube-api-access-6f27m\") pod \"d843db5d-75e6-4f95-a297-41af3451648f\" (UID: \"d843db5d-75e6-4f95-a297-41af3451648f\") " Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.076085 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d843db5d-75e6-4f95-a297-41af3451648f-kube-api-access-6f27m" (OuterVolumeSpecName: "kube-api-access-6f27m") pod "d843db5d-75e6-4f95-a297-41af3451648f" (UID: "d843db5d-75e6-4f95-a297-41af3451648f"). InnerVolumeSpecName "kube-api-access-6f27m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.079580 4959 scope.go:117] "RemoveContainer" containerID="e5664b833d564c90e189d17aa8f318a3e302fb3e3811e034b2ade020fc05e777" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.096942 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d843db5d-75e6-4f95-a297-41af3451648f-inventory" (OuterVolumeSpecName: "inventory") pod "d843db5d-75e6-4f95-a297-41af3451648f" (UID: "d843db5d-75e6-4f95-a297-41af3451648f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.098295 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d843db5d-75e6-4f95-a297-41af3451648f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d843db5d-75e6-4f95-a297-41af3451648f" (UID: "d843db5d-75e6-4f95-a297-41af3451648f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.173628 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d843db5d-75e6-4f95-a297-41af3451648f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.173686 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6f27m\" (UniqueName: \"kubernetes.io/projected/d843db5d-75e6-4f95-a297-41af3451648f-kube-api-access-6f27m\") on node \"crc\" DevicePath \"\"" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.173707 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d843db5d-75e6-4f95-a297-41af3451648f-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.658971 4959 scope.go:117] "RemoveContainer" containerID="8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1" Oct 07 14:12:28 crc kubenswrapper[4959]: E1007 14:12:28.659666 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.683894 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q" event={"ID":"d843db5d-75e6-4f95-a297-41af3451648f","Type":"ContainerDied","Data":"f2fa062db5a3a7e3bb28c9874850decf41d5affa6611dfa06d9eef3e02c21f1e"} Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.683960 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.683968 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2fa062db5a3a7e3bb28c9874850decf41d5affa6611dfa06d9eef3e02c21f1e" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.770449 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k"] Oct 07 14:12:28 crc kubenswrapper[4959]: E1007 14:12:28.770779 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63dd100a-8db8-4adb-810c-dc2c06a7371a" containerName="extract-content" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.770790 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="63dd100a-8db8-4adb-810c-dc2c06a7371a" containerName="extract-content" Oct 07 14:12:28 crc kubenswrapper[4959]: E1007 14:12:28.770807 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d843db5d-75e6-4f95-a297-41af3451648f" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.770815 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d843db5d-75e6-4f95-a297-41af3451648f" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 07 14:12:28 crc kubenswrapper[4959]: E1007 14:12:28.770840 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63dd100a-8db8-4adb-810c-dc2c06a7371a" containerName="registry-server" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.770845 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="63dd100a-8db8-4adb-810c-dc2c06a7371a" containerName="registry-server" Oct 07 14:12:28 crc kubenswrapper[4959]: E1007 14:12:28.770855 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63dd100a-8db8-4adb-810c-dc2c06a7371a" containerName="extract-utilities" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.770861 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="63dd100a-8db8-4adb-810c-dc2c06a7371a" containerName="extract-utilities" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.771043 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="63dd100a-8db8-4adb-810c-dc2c06a7371a" containerName="registry-server" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.771064 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="d843db5d-75e6-4f95-a297-41af3451648f" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.771655 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.777628 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.777768 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.777873 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-c6vmh" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.777975 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.782912 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k"] Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.885658 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95tr9\" (UniqueName: \"kubernetes.io/projected/fead8dac-8646-4828-be8e-d17bc43d7d61-kube-api-access-95tr9\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k\" (UID: \"fead8dac-8646-4828-be8e-d17bc43d7d61\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.885946 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fead8dac-8646-4828-be8e-d17bc43d7d61-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k\" (UID: \"fead8dac-8646-4828-be8e-d17bc43d7d61\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.886159 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fead8dac-8646-4828-be8e-d17bc43d7d61-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k\" (UID: \"fead8dac-8646-4828-be8e-d17bc43d7d61\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.988320 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95tr9\" (UniqueName: \"kubernetes.io/projected/fead8dac-8646-4828-be8e-d17bc43d7d61-kube-api-access-95tr9\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k\" (UID: \"fead8dac-8646-4828-be8e-d17bc43d7d61\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.988464 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fead8dac-8646-4828-be8e-d17bc43d7d61-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k\" (UID: \"fead8dac-8646-4828-be8e-d17bc43d7d61\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.988492 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fead8dac-8646-4828-be8e-d17bc43d7d61-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k\" (UID: \"fead8dac-8646-4828-be8e-d17bc43d7d61\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k" Oct 07 14:12:28 crc kubenswrapper[4959]: I1007 14:12:28.995945 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fead8dac-8646-4828-be8e-d17bc43d7d61-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k\" (UID: \"fead8dac-8646-4828-be8e-d17bc43d7d61\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k" Oct 07 14:12:29 crc kubenswrapper[4959]: I1007 14:12:29.008873 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fead8dac-8646-4828-be8e-d17bc43d7d61-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k\" (UID: \"fead8dac-8646-4828-be8e-d17bc43d7d61\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k" Oct 07 14:12:29 crc kubenswrapper[4959]: I1007 14:12:29.015232 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95tr9\" (UniqueName: \"kubernetes.io/projected/fead8dac-8646-4828-be8e-d17bc43d7d61-kube-api-access-95tr9\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k\" (UID: \"fead8dac-8646-4828-be8e-d17bc43d7d61\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k" Oct 07 14:12:29 crc kubenswrapper[4959]: I1007 14:12:29.164904 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k" Oct 07 14:12:29 crc kubenswrapper[4959]: I1007 14:12:29.686467 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k"] Oct 07 14:12:30 crc kubenswrapper[4959]: I1007 14:12:30.701045 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k" event={"ID":"fead8dac-8646-4828-be8e-d17bc43d7d61","Type":"ContainerStarted","Data":"5c0eda4a3d50c3d8103a5febc1d8b9779b030de9e385b63bbb1f48ab66c22544"} Oct 07 14:12:31 crc kubenswrapper[4959]: I1007 14:12:31.709442 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k" event={"ID":"fead8dac-8646-4828-be8e-d17bc43d7d61","Type":"ContainerStarted","Data":"8c2b3d835cf81c27fb72a2add3653dfa9fd1caf6bb28054442b2e29a6e0f8742"} Oct 07 14:12:31 crc kubenswrapper[4959]: I1007 14:12:31.731752 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k" podStartSLOduration=2.629704294 podStartE2EDuration="3.731731749s" podCreationTimestamp="2025-10-07 14:12:28 +0000 UTC" firstStartedPulling="2025-10-07 14:12:29.689883562 +0000 UTC m=+1631.773287897" lastFinishedPulling="2025-10-07 14:12:30.791911027 +0000 UTC m=+1632.875315352" observedRunningTime="2025-10-07 14:12:31.725165436 +0000 UTC m=+1633.808569791" watchObservedRunningTime="2025-10-07 14:12:31.731731749 +0000 UTC m=+1633.815136064" Oct 07 14:12:32 crc kubenswrapper[4959]: I1007 14:12:32.026156 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-gvngb"] Oct 07 14:12:32 crc kubenswrapper[4959]: I1007 14:12:32.033219 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-gvngb"] Oct 07 14:12:32 crc kubenswrapper[4959]: I1007 14:12:32.663577 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f" path="/var/lib/kubelet/pods/c0d4aa0d-e280-42c3-8a49-6dcd6105ab6f/volumes" Oct 07 14:12:34 crc kubenswrapper[4959]: I1007 14:12:34.035120 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-fjrkz"] Oct 07 14:12:34 crc kubenswrapper[4959]: I1007 14:12:34.042903 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-3409-account-create-fwppc"] Oct 07 14:12:34 crc kubenswrapper[4959]: I1007 14:12:34.053066 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-d404-account-create-w2qcd"] Oct 07 14:12:34 crc kubenswrapper[4959]: I1007 14:12:34.059296 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-3d71-account-create-tpxcs"] Oct 07 14:12:34 crc kubenswrapper[4959]: I1007 14:12:34.065366 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-fjrkz"] Oct 07 14:12:34 crc kubenswrapper[4959]: I1007 14:12:34.071883 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-3409-account-create-fwppc"] Oct 07 14:12:34 crc kubenswrapper[4959]: I1007 14:12:34.078255 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-3d71-account-create-tpxcs"] Oct 07 14:12:34 crc kubenswrapper[4959]: I1007 14:12:34.084478 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-d404-account-create-w2qcd"] Oct 07 14:12:34 crc kubenswrapper[4959]: I1007 14:12:34.664963 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="024f57ee-b5d4-4f45-b447-5410cf1722ec" path="/var/lib/kubelet/pods/024f57ee-b5d4-4f45-b447-5410cf1722ec/volumes" Oct 07 14:12:34 crc kubenswrapper[4959]: I1007 14:12:34.665634 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56568d60-1394-4d21-af2c-2b5fa6bf8007" path="/var/lib/kubelet/pods/56568d60-1394-4d21-af2c-2b5fa6bf8007/volumes" Oct 07 14:12:34 crc kubenswrapper[4959]: I1007 14:12:34.666128 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78ce8105-cd45-40fd-bfcf-2d853c8d822a" path="/var/lib/kubelet/pods/78ce8105-cd45-40fd-bfcf-2d853c8d822a/volumes" Oct 07 14:12:34 crc kubenswrapper[4959]: I1007 14:12:34.666589 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c88deaae-f6a8-45c8-b17b-6dc226c4f3d6" path="/var/lib/kubelet/pods/c88deaae-f6a8-45c8-b17b-6dc226c4f3d6/volumes" Oct 07 14:12:35 crc kubenswrapper[4959]: I1007 14:12:35.743690 4959 generic.go:334] "Generic (PLEG): container finished" podID="fead8dac-8646-4828-be8e-d17bc43d7d61" containerID="8c2b3d835cf81c27fb72a2add3653dfa9fd1caf6bb28054442b2e29a6e0f8742" exitCode=0 Oct 07 14:12:35 crc kubenswrapper[4959]: I1007 14:12:35.743868 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k" event={"ID":"fead8dac-8646-4828-be8e-d17bc43d7d61","Type":"ContainerDied","Data":"8c2b3d835cf81c27fb72a2add3653dfa9fd1caf6bb28054442b2e29a6e0f8742"} Oct 07 14:12:37 crc kubenswrapper[4959]: I1007 14:12:37.211194 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k" Oct 07 14:12:37 crc kubenswrapper[4959]: I1007 14:12:37.373972 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fead8dac-8646-4828-be8e-d17bc43d7d61-inventory\") pod \"fead8dac-8646-4828-be8e-d17bc43d7d61\" (UID: \"fead8dac-8646-4828-be8e-d17bc43d7d61\") " Oct 07 14:12:37 crc kubenswrapper[4959]: I1007 14:12:37.374230 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95tr9\" (UniqueName: \"kubernetes.io/projected/fead8dac-8646-4828-be8e-d17bc43d7d61-kube-api-access-95tr9\") pod \"fead8dac-8646-4828-be8e-d17bc43d7d61\" (UID: \"fead8dac-8646-4828-be8e-d17bc43d7d61\") " Oct 07 14:12:37 crc kubenswrapper[4959]: I1007 14:12:37.374314 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fead8dac-8646-4828-be8e-d17bc43d7d61-ssh-key\") pod \"fead8dac-8646-4828-be8e-d17bc43d7d61\" (UID: \"fead8dac-8646-4828-be8e-d17bc43d7d61\") " Oct 07 14:12:37 crc kubenswrapper[4959]: I1007 14:12:37.382394 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fead8dac-8646-4828-be8e-d17bc43d7d61-kube-api-access-95tr9" (OuterVolumeSpecName: "kube-api-access-95tr9") pod "fead8dac-8646-4828-be8e-d17bc43d7d61" (UID: "fead8dac-8646-4828-be8e-d17bc43d7d61"). InnerVolumeSpecName "kube-api-access-95tr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:12:37 crc kubenswrapper[4959]: I1007 14:12:37.397989 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fead8dac-8646-4828-be8e-d17bc43d7d61-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fead8dac-8646-4828-be8e-d17bc43d7d61" (UID: "fead8dac-8646-4828-be8e-d17bc43d7d61"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:12:37 crc kubenswrapper[4959]: I1007 14:12:37.420721 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fead8dac-8646-4828-be8e-d17bc43d7d61-inventory" (OuterVolumeSpecName: "inventory") pod "fead8dac-8646-4828-be8e-d17bc43d7d61" (UID: "fead8dac-8646-4828-be8e-d17bc43d7d61"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:12:37 crc kubenswrapper[4959]: I1007 14:12:37.476478 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fead8dac-8646-4828-be8e-d17bc43d7d61-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 14:12:37 crc kubenswrapper[4959]: I1007 14:12:37.476518 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95tr9\" (UniqueName: \"kubernetes.io/projected/fead8dac-8646-4828-be8e-d17bc43d7d61-kube-api-access-95tr9\") on node \"crc\" DevicePath \"\"" Oct 07 14:12:37 crc kubenswrapper[4959]: I1007 14:12:37.476528 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fead8dac-8646-4828-be8e-d17bc43d7d61-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 14:12:37 crc kubenswrapper[4959]: I1007 14:12:37.768586 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k" event={"ID":"fead8dac-8646-4828-be8e-d17bc43d7d61","Type":"ContainerDied","Data":"5c0eda4a3d50c3d8103a5febc1d8b9779b030de9e385b63bbb1f48ab66c22544"} Oct 07 14:12:37 crc kubenswrapper[4959]: I1007 14:12:37.768957 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c0eda4a3d50c3d8103a5febc1d8b9779b030de9e385b63bbb1f48ab66c22544" Oct 07 14:12:37 crc kubenswrapper[4959]: I1007 14:12:37.768644 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k" Oct 07 14:12:37 crc kubenswrapper[4959]: I1007 14:12:37.843034 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-wjl8b"] Oct 07 14:12:37 crc kubenswrapper[4959]: E1007 14:12:37.843543 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fead8dac-8646-4828-be8e-d17bc43d7d61" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 07 14:12:37 crc kubenswrapper[4959]: I1007 14:12:37.843567 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="fead8dac-8646-4828-be8e-d17bc43d7d61" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 07 14:12:37 crc kubenswrapper[4959]: I1007 14:12:37.843819 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="fead8dac-8646-4828-be8e-d17bc43d7d61" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 07 14:12:37 crc kubenswrapper[4959]: I1007 14:12:37.844568 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wjl8b" Oct 07 14:12:37 crc kubenswrapper[4959]: I1007 14:12:37.846286 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 14:12:37 crc kubenswrapper[4959]: I1007 14:12:37.846628 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-c6vmh" Oct 07 14:12:37 crc kubenswrapper[4959]: I1007 14:12:37.846914 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 14:12:37 crc kubenswrapper[4959]: I1007 14:12:37.849701 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 14:12:37 crc kubenswrapper[4959]: I1007 14:12:37.852486 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-wjl8b"] Oct 07 14:12:37 crc kubenswrapper[4959]: I1007 14:12:37.984340 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xbsd\" (UniqueName: \"kubernetes.io/projected/e2de54f5-c547-4a2c-afa3-2ba5715d423b-kube-api-access-2xbsd\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-wjl8b\" (UID: \"e2de54f5-c547-4a2c-afa3-2ba5715d423b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wjl8b" Oct 07 14:12:37 crc kubenswrapper[4959]: I1007 14:12:37.984600 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e2de54f5-c547-4a2c-afa3-2ba5715d423b-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-wjl8b\" (UID: \"e2de54f5-c547-4a2c-afa3-2ba5715d423b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wjl8b" Oct 07 14:12:37 crc kubenswrapper[4959]: I1007 14:12:37.984776 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2de54f5-c547-4a2c-afa3-2ba5715d423b-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-wjl8b\" (UID: \"e2de54f5-c547-4a2c-afa3-2ba5715d423b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wjl8b" Oct 07 14:12:38 crc kubenswrapper[4959]: I1007 14:12:38.085989 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xbsd\" (UniqueName: \"kubernetes.io/projected/e2de54f5-c547-4a2c-afa3-2ba5715d423b-kube-api-access-2xbsd\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-wjl8b\" (UID: \"e2de54f5-c547-4a2c-afa3-2ba5715d423b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wjl8b" Oct 07 14:12:38 crc kubenswrapper[4959]: I1007 14:12:38.086062 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e2de54f5-c547-4a2c-afa3-2ba5715d423b-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-wjl8b\" (UID: \"e2de54f5-c547-4a2c-afa3-2ba5715d423b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wjl8b" Oct 07 14:12:38 crc kubenswrapper[4959]: I1007 14:12:38.086121 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2de54f5-c547-4a2c-afa3-2ba5715d423b-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-wjl8b\" (UID: \"e2de54f5-c547-4a2c-afa3-2ba5715d423b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wjl8b" Oct 07 14:12:38 crc kubenswrapper[4959]: I1007 14:12:38.089945 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2de54f5-c547-4a2c-afa3-2ba5715d423b-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-wjl8b\" (UID: \"e2de54f5-c547-4a2c-afa3-2ba5715d423b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wjl8b" Oct 07 14:12:38 crc kubenswrapper[4959]: I1007 14:12:38.090731 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e2de54f5-c547-4a2c-afa3-2ba5715d423b-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-wjl8b\" (UID: \"e2de54f5-c547-4a2c-afa3-2ba5715d423b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wjl8b" Oct 07 14:12:38 crc kubenswrapper[4959]: I1007 14:12:38.106485 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xbsd\" (UniqueName: \"kubernetes.io/projected/e2de54f5-c547-4a2c-afa3-2ba5715d423b-kube-api-access-2xbsd\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-wjl8b\" (UID: \"e2de54f5-c547-4a2c-afa3-2ba5715d423b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wjl8b" Oct 07 14:12:38 crc kubenswrapper[4959]: I1007 14:12:38.164202 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wjl8b" Oct 07 14:12:38 crc kubenswrapper[4959]: I1007 14:12:38.708090 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-wjl8b"] Oct 07 14:12:38 crc kubenswrapper[4959]: I1007 14:12:38.779919 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wjl8b" event={"ID":"e2de54f5-c547-4a2c-afa3-2ba5715d423b","Type":"ContainerStarted","Data":"fe4ca5a8da9ac8a861fdf479b82ecf6122fee8f6a583b4e6eedce215182505ab"} Oct 07 14:12:39 crc kubenswrapper[4959]: I1007 14:12:39.789271 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wjl8b" event={"ID":"e2de54f5-c547-4a2c-afa3-2ba5715d423b","Type":"ContainerStarted","Data":"aefc8d2781df2b525eacaaa3711fe17975d192444d61361f987caf6586ee2a0d"} Oct 07 14:12:39 crc kubenswrapper[4959]: I1007 14:12:39.807929 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wjl8b" podStartSLOduration=2.199481632 podStartE2EDuration="2.807909916s" podCreationTimestamp="2025-10-07 14:12:37 +0000 UTC" firstStartedPulling="2025-10-07 14:12:38.726007892 +0000 UTC m=+1640.809412217" lastFinishedPulling="2025-10-07 14:12:39.334436186 +0000 UTC m=+1641.417840501" observedRunningTime="2025-10-07 14:12:39.807293741 +0000 UTC m=+1641.890698066" watchObservedRunningTime="2025-10-07 14:12:39.807909916 +0000 UTC m=+1641.891314231" Oct 07 14:12:42 crc kubenswrapper[4959]: I1007 14:12:42.656764 4959 scope.go:117] "RemoveContainer" containerID="8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1" Oct 07 14:12:42 crc kubenswrapper[4959]: E1007 14:12:42.657940 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:12:57 crc kubenswrapper[4959]: I1007 14:12:57.653919 4959 scope.go:117] "RemoveContainer" containerID="8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1" Oct 07 14:12:57 crc kubenswrapper[4959]: E1007 14:12:57.655091 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:13:10 crc kubenswrapper[4959]: I1007 14:13:10.654189 4959 scope.go:117] "RemoveContainer" containerID="8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1" Oct 07 14:13:10 crc kubenswrapper[4959]: E1007 14:13:10.655092 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:13:16 crc kubenswrapper[4959]: I1007 14:13:16.190962 4959 generic.go:334] "Generic (PLEG): container finished" podID="e2de54f5-c547-4a2c-afa3-2ba5715d423b" containerID="aefc8d2781df2b525eacaaa3711fe17975d192444d61361f987caf6586ee2a0d" exitCode=0 Oct 07 14:13:16 crc kubenswrapper[4959]: I1007 14:13:16.191026 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wjl8b" event={"ID":"e2de54f5-c547-4a2c-afa3-2ba5715d423b","Type":"ContainerDied","Data":"aefc8d2781df2b525eacaaa3711fe17975d192444d61361f987caf6586ee2a0d"} Oct 07 14:13:17 crc kubenswrapper[4959]: I1007 14:13:17.611703 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wjl8b" Oct 07 14:13:17 crc kubenswrapper[4959]: I1007 14:13:17.691043 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xbsd\" (UniqueName: \"kubernetes.io/projected/e2de54f5-c547-4a2c-afa3-2ba5715d423b-kube-api-access-2xbsd\") pod \"e2de54f5-c547-4a2c-afa3-2ba5715d423b\" (UID: \"e2de54f5-c547-4a2c-afa3-2ba5715d423b\") " Oct 07 14:13:17 crc kubenswrapper[4959]: I1007 14:13:17.691743 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e2de54f5-c547-4a2c-afa3-2ba5715d423b-ssh-key\") pod \"e2de54f5-c547-4a2c-afa3-2ba5715d423b\" (UID: \"e2de54f5-c547-4a2c-afa3-2ba5715d423b\") " Oct 07 14:13:17 crc kubenswrapper[4959]: I1007 14:13:17.692122 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2de54f5-c547-4a2c-afa3-2ba5715d423b-inventory\") pod \"e2de54f5-c547-4a2c-afa3-2ba5715d423b\" (UID: \"e2de54f5-c547-4a2c-afa3-2ba5715d423b\") " Oct 07 14:13:17 crc kubenswrapper[4959]: I1007 14:13:17.698732 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2de54f5-c547-4a2c-afa3-2ba5715d423b-kube-api-access-2xbsd" (OuterVolumeSpecName: "kube-api-access-2xbsd") pod "e2de54f5-c547-4a2c-afa3-2ba5715d423b" (UID: "e2de54f5-c547-4a2c-afa3-2ba5715d423b"). InnerVolumeSpecName "kube-api-access-2xbsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:13:17 crc kubenswrapper[4959]: I1007 14:13:17.724848 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2de54f5-c547-4a2c-afa3-2ba5715d423b-inventory" (OuterVolumeSpecName: "inventory") pod "e2de54f5-c547-4a2c-afa3-2ba5715d423b" (UID: "e2de54f5-c547-4a2c-afa3-2ba5715d423b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:13:17 crc kubenswrapper[4959]: I1007 14:13:17.725731 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2de54f5-c547-4a2c-afa3-2ba5715d423b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e2de54f5-c547-4a2c-afa3-2ba5715d423b" (UID: "e2de54f5-c547-4a2c-afa3-2ba5715d423b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:13:17 crc kubenswrapper[4959]: I1007 14:13:17.794188 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2de54f5-c547-4a2c-afa3-2ba5715d423b-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 14:13:17 crc kubenswrapper[4959]: I1007 14:13:17.794224 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xbsd\" (UniqueName: \"kubernetes.io/projected/e2de54f5-c547-4a2c-afa3-2ba5715d423b-kube-api-access-2xbsd\") on node \"crc\" DevicePath \"\"" Oct 07 14:13:17 crc kubenswrapper[4959]: I1007 14:13:17.794236 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e2de54f5-c547-4a2c-afa3-2ba5715d423b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 14:13:18 crc kubenswrapper[4959]: I1007 14:13:18.212083 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wjl8b" event={"ID":"e2de54f5-c547-4a2c-afa3-2ba5715d423b","Type":"ContainerDied","Data":"fe4ca5a8da9ac8a861fdf479b82ecf6122fee8f6a583b4e6eedce215182505ab"} Oct 07 14:13:18 crc kubenswrapper[4959]: I1007 14:13:18.212209 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe4ca5a8da9ac8a861fdf479b82ecf6122fee8f6a583b4e6eedce215182505ab" Oct 07 14:13:18 crc kubenswrapper[4959]: I1007 14:13:18.212210 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wjl8b" Oct 07 14:13:18 crc kubenswrapper[4959]: I1007 14:13:18.290651 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg"] Oct 07 14:13:18 crc kubenswrapper[4959]: E1007 14:13:18.291493 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2de54f5-c547-4a2c-afa3-2ba5715d423b" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 07 14:13:18 crc kubenswrapper[4959]: I1007 14:13:18.291515 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2de54f5-c547-4a2c-afa3-2ba5715d423b" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 07 14:13:18 crc kubenswrapper[4959]: I1007 14:13:18.291728 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2de54f5-c547-4a2c-afa3-2ba5715d423b" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 07 14:13:18 crc kubenswrapper[4959]: I1007 14:13:18.292640 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg" Oct 07 14:13:18 crc kubenswrapper[4959]: I1007 14:13:18.300411 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 14:13:18 crc kubenswrapper[4959]: I1007 14:13:18.300953 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 14:13:18 crc kubenswrapper[4959]: I1007 14:13:18.303993 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-c6vmh" Oct 07 14:13:18 crc kubenswrapper[4959]: I1007 14:13:18.304474 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 14:13:18 crc kubenswrapper[4959]: I1007 14:13:18.334771 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg"] Oct 07 14:13:18 crc kubenswrapper[4959]: I1007 14:13:18.423317 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d983b0f4-2357-46c9-bd51-32efc75ec771-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg\" (UID: \"d983b0f4-2357-46c9-bd51-32efc75ec771\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg" Oct 07 14:13:18 crc kubenswrapper[4959]: I1007 14:13:18.423646 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d983b0f4-2357-46c9-bd51-32efc75ec771-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg\" (UID: \"d983b0f4-2357-46c9-bd51-32efc75ec771\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg" Oct 07 14:13:18 crc kubenswrapper[4959]: I1007 14:13:18.424390 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9jzs\" (UniqueName: \"kubernetes.io/projected/d983b0f4-2357-46c9-bd51-32efc75ec771-kube-api-access-g9jzs\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg\" (UID: \"d983b0f4-2357-46c9-bd51-32efc75ec771\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg" Oct 07 14:13:18 crc kubenswrapper[4959]: I1007 14:13:18.526731 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9jzs\" (UniqueName: \"kubernetes.io/projected/d983b0f4-2357-46c9-bd51-32efc75ec771-kube-api-access-g9jzs\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg\" (UID: \"d983b0f4-2357-46c9-bd51-32efc75ec771\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg" Oct 07 14:13:18 crc kubenswrapper[4959]: I1007 14:13:18.526817 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d983b0f4-2357-46c9-bd51-32efc75ec771-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg\" (UID: \"d983b0f4-2357-46c9-bd51-32efc75ec771\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg" Oct 07 14:13:18 crc kubenswrapper[4959]: I1007 14:13:18.526848 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d983b0f4-2357-46c9-bd51-32efc75ec771-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg\" (UID: \"d983b0f4-2357-46c9-bd51-32efc75ec771\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg" Oct 07 14:13:18 crc kubenswrapper[4959]: I1007 14:13:18.530896 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d983b0f4-2357-46c9-bd51-32efc75ec771-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg\" (UID: \"d983b0f4-2357-46c9-bd51-32efc75ec771\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg" Oct 07 14:13:18 crc kubenswrapper[4959]: I1007 14:13:18.534225 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d983b0f4-2357-46c9-bd51-32efc75ec771-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg\" (UID: \"d983b0f4-2357-46c9-bd51-32efc75ec771\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg" Oct 07 14:13:18 crc kubenswrapper[4959]: I1007 14:13:18.553515 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9jzs\" (UniqueName: \"kubernetes.io/projected/d983b0f4-2357-46c9-bd51-32efc75ec771-kube-api-access-g9jzs\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg\" (UID: \"d983b0f4-2357-46c9-bd51-32efc75ec771\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg" Oct 07 14:13:18 crc kubenswrapper[4959]: I1007 14:13:18.619019 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-c6vmh" Oct 07 14:13:18 crc kubenswrapper[4959]: I1007 14:13:18.626652 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg" Oct 07 14:13:19 crc kubenswrapper[4959]: I1007 14:13:19.173777 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg"] Oct 07 14:13:19 crc kubenswrapper[4959]: I1007 14:13:19.227846 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg" event={"ID":"d983b0f4-2357-46c9-bd51-32efc75ec771","Type":"ContainerStarted","Data":"a11ea4b2cef04615c4ea7bda6b953da3da58d6973f9fdba7fe05d7d074cf564f"} Oct 07 14:13:19 crc kubenswrapper[4959]: I1007 14:13:19.974720 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 14:13:21 crc kubenswrapper[4959]: I1007 14:13:21.247963 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg" event={"ID":"d983b0f4-2357-46c9-bd51-32efc75ec771","Type":"ContainerStarted","Data":"412050b987b86adccfa498c30b05e722afed7366b18779c3715eeae0cfd661a8"} Oct 07 14:13:21 crc kubenswrapper[4959]: I1007 14:13:21.280753 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg" podStartSLOduration=2.484142074 podStartE2EDuration="3.280727605s" podCreationTimestamp="2025-10-07 14:13:18 +0000 UTC" firstStartedPulling="2025-10-07 14:13:19.173812477 +0000 UTC m=+1681.257216802" lastFinishedPulling="2025-10-07 14:13:19.970398008 +0000 UTC m=+1682.053802333" observedRunningTime="2025-10-07 14:13:21.268454589 +0000 UTC m=+1683.351858914" watchObservedRunningTime="2025-10-07 14:13:21.280727605 +0000 UTC m=+1683.364131930" Oct 07 14:13:23 crc kubenswrapper[4959]: I1007 14:13:23.653178 4959 scope.go:117] "RemoveContainer" containerID="8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1" Oct 07 14:13:23 crc kubenswrapper[4959]: E1007 14:13:23.654416 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:13:25 crc kubenswrapper[4959]: I1007 14:13:25.289983 4959 generic.go:334] "Generic (PLEG): container finished" podID="d983b0f4-2357-46c9-bd51-32efc75ec771" containerID="412050b987b86adccfa498c30b05e722afed7366b18779c3715eeae0cfd661a8" exitCode=0 Oct 07 14:13:25 crc kubenswrapper[4959]: I1007 14:13:25.290032 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg" event={"ID":"d983b0f4-2357-46c9-bd51-32efc75ec771","Type":"ContainerDied","Data":"412050b987b86adccfa498c30b05e722afed7366b18779c3715eeae0cfd661a8"} Oct 07 14:13:26 crc kubenswrapper[4959]: I1007 14:13:26.707509 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg" Oct 07 14:13:26 crc kubenswrapper[4959]: I1007 14:13:26.799155 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d983b0f4-2357-46c9-bd51-32efc75ec771-inventory\") pod \"d983b0f4-2357-46c9-bd51-32efc75ec771\" (UID: \"d983b0f4-2357-46c9-bd51-32efc75ec771\") " Oct 07 14:13:26 crc kubenswrapper[4959]: I1007 14:13:26.799286 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d983b0f4-2357-46c9-bd51-32efc75ec771-ssh-key\") pod \"d983b0f4-2357-46c9-bd51-32efc75ec771\" (UID: \"d983b0f4-2357-46c9-bd51-32efc75ec771\") " Oct 07 14:13:26 crc kubenswrapper[4959]: I1007 14:13:26.799390 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9jzs\" (UniqueName: \"kubernetes.io/projected/d983b0f4-2357-46c9-bd51-32efc75ec771-kube-api-access-g9jzs\") pod \"d983b0f4-2357-46c9-bd51-32efc75ec771\" (UID: \"d983b0f4-2357-46c9-bd51-32efc75ec771\") " Oct 07 14:13:26 crc kubenswrapper[4959]: I1007 14:13:26.804502 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d983b0f4-2357-46c9-bd51-32efc75ec771-kube-api-access-g9jzs" (OuterVolumeSpecName: "kube-api-access-g9jzs") pod "d983b0f4-2357-46c9-bd51-32efc75ec771" (UID: "d983b0f4-2357-46c9-bd51-32efc75ec771"). InnerVolumeSpecName "kube-api-access-g9jzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:13:26 crc kubenswrapper[4959]: I1007 14:13:26.825307 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d983b0f4-2357-46c9-bd51-32efc75ec771-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d983b0f4-2357-46c9-bd51-32efc75ec771" (UID: "d983b0f4-2357-46c9-bd51-32efc75ec771"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:13:26 crc kubenswrapper[4959]: I1007 14:13:26.826868 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d983b0f4-2357-46c9-bd51-32efc75ec771-inventory" (OuterVolumeSpecName: "inventory") pod "d983b0f4-2357-46c9-bd51-32efc75ec771" (UID: "d983b0f4-2357-46c9-bd51-32efc75ec771"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:13:26 crc kubenswrapper[4959]: I1007 14:13:26.901533 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9jzs\" (UniqueName: \"kubernetes.io/projected/d983b0f4-2357-46c9-bd51-32efc75ec771-kube-api-access-g9jzs\") on node \"crc\" DevicePath \"\"" Oct 07 14:13:26 crc kubenswrapper[4959]: I1007 14:13:26.901568 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d983b0f4-2357-46c9-bd51-32efc75ec771-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 14:13:26 crc kubenswrapper[4959]: I1007 14:13:26.901579 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d983b0f4-2357-46c9-bd51-32efc75ec771-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 14:13:27 crc kubenswrapper[4959]: I1007 14:13:27.308870 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg" event={"ID":"d983b0f4-2357-46c9-bd51-32efc75ec771","Type":"ContainerDied","Data":"a11ea4b2cef04615c4ea7bda6b953da3da58d6973f9fdba7fe05d7d074cf564f"} Oct 07 14:13:27 crc kubenswrapper[4959]: I1007 14:13:27.308923 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a11ea4b2cef04615c4ea7bda6b953da3da58d6973f9fdba7fe05d7d074cf564f" Oct 07 14:13:27 crc kubenswrapper[4959]: I1007 14:13:27.308978 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg" Oct 07 14:13:27 crc kubenswrapper[4959]: I1007 14:13:27.381377 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc"] Oct 07 14:13:27 crc kubenswrapper[4959]: E1007 14:13:27.381831 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d983b0f4-2357-46c9-bd51-32efc75ec771" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 07 14:13:27 crc kubenswrapper[4959]: I1007 14:13:27.381853 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d983b0f4-2357-46c9-bd51-32efc75ec771" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 07 14:13:27 crc kubenswrapper[4959]: I1007 14:13:27.382103 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="d983b0f4-2357-46c9-bd51-32efc75ec771" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 07 14:13:27 crc kubenswrapper[4959]: I1007 14:13:27.382874 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc" Oct 07 14:13:27 crc kubenswrapper[4959]: I1007 14:13:27.384919 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 14:13:27 crc kubenswrapper[4959]: I1007 14:13:27.385317 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 14:13:27 crc kubenswrapper[4959]: I1007 14:13:27.385490 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 14:13:27 crc kubenswrapper[4959]: I1007 14:13:27.385657 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-c6vmh" Oct 07 14:13:27 crc kubenswrapper[4959]: I1007 14:13:27.409619 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc"] Oct 07 14:13:27 crc kubenswrapper[4959]: I1007 14:13:27.511740 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gmf2\" (UniqueName: \"kubernetes.io/projected/4d310c8e-5f3a-484f-ab93-ee382d288df8-kube-api-access-8gmf2\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc\" (UID: \"4d310c8e-5f3a-484f-ab93-ee382d288df8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc" Oct 07 14:13:27 crc kubenswrapper[4959]: I1007 14:13:27.511804 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d310c8e-5f3a-484f-ab93-ee382d288df8-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc\" (UID: \"4d310c8e-5f3a-484f-ab93-ee382d288df8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc" Oct 07 14:13:27 crc kubenswrapper[4959]: I1007 14:13:27.512043 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d310c8e-5f3a-484f-ab93-ee382d288df8-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc\" (UID: \"4d310c8e-5f3a-484f-ab93-ee382d288df8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc" Oct 07 14:13:27 crc kubenswrapper[4959]: I1007 14:13:27.613568 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d310c8e-5f3a-484f-ab93-ee382d288df8-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc\" (UID: \"4d310c8e-5f3a-484f-ab93-ee382d288df8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc" Oct 07 14:13:27 crc kubenswrapper[4959]: I1007 14:13:27.613749 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gmf2\" (UniqueName: \"kubernetes.io/projected/4d310c8e-5f3a-484f-ab93-ee382d288df8-kube-api-access-8gmf2\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc\" (UID: \"4d310c8e-5f3a-484f-ab93-ee382d288df8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc" Oct 07 14:13:27 crc kubenswrapper[4959]: I1007 14:13:27.613799 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d310c8e-5f3a-484f-ab93-ee382d288df8-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc\" (UID: \"4d310c8e-5f3a-484f-ab93-ee382d288df8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc" Oct 07 14:13:27 crc kubenswrapper[4959]: I1007 14:13:27.617476 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d310c8e-5f3a-484f-ab93-ee382d288df8-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc\" (UID: \"4d310c8e-5f3a-484f-ab93-ee382d288df8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc" Oct 07 14:13:27 crc kubenswrapper[4959]: I1007 14:13:27.619994 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d310c8e-5f3a-484f-ab93-ee382d288df8-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc\" (UID: \"4d310c8e-5f3a-484f-ab93-ee382d288df8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc" Oct 07 14:13:27 crc kubenswrapper[4959]: I1007 14:13:27.639095 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gmf2\" (UniqueName: \"kubernetes.io/projected/4d310c8e-5f3a-484f-ab93-ee382d288df8-kube-api-access-8gmf2\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc\" (UID: \"4d310c8e-5f3a-484f-ab93-ee382d288df8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc" Oct 07 14:13:27 crc kubenswrapper[4959]: I1007 14:13:27.703773 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc" Oct 07 14:13:28 crc kubenswrapper[4959]: I1007 14:13:28.222771 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc"] Oct 07 14:13:28 crc kubenswrapper[4959]: W1007 14:13:28.227021 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d310c8e_5f3a_484f_ab93_ee382d288df8.slice/crio-938d12d11a6f1ae863bd84d588c222cd85cca4965d8283822d12e183b52d9dd2 WatchSource:0}: Error finding container 938d12d11a6f1ae863bd84d588c222cd85cca4965d8283822d12e183b52d9dd2: Status 404 returned error can't find the container with id 938d12d11a6f1ae863bd84d588c222cd85cca4965d8283822d12e183b52d9dd2 Oct 07 14:13:28 crc kubenswrapper[4959]: I1007 14:13:28.293593 4959 scope.go:117] "RemoveContainer" containerID="33bc5d4e608c9f7275ee45befb6853744451f5dffcc0951284c7709cfaced467" Oct 07 14:13:28 crc kubenswrapper[4959]: I1007 14:13:28.320870 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc" event={"ID":"4d310c8e-5f3a-484f-ab93-ee382d288df8","Type":"ContainerStarted","Data":"938d12d11a6f1ae863bd84d588c222cd85cca4965d8283822d12e183b52d9dd2"} Oct 07 14:13:28 crc kubenswrapper[4959]: I1007 14:13:28.330460 4959 scope.go:117] "RemoveContainer" containerID="a20ca8a566174703ad0d5a83c6ea9b5dd928950691431146eba300ba5a1de4ee" Oct 07 14:13:28 crc kubenswrapper[4959]: I1007 14:13:28.350397 4959 scope.go:117] "RemoveContainer" containerID="7e6ad7d9a1a80044b01306c65e9e6b1491b2359f709a14cc816497272c376c9b" Oct 07 14:13:28 crc kubenswrapper[4959]: I1007 14:13:28.368619 4959 scope.go:117] "RemoveContainer" containerID="4e442444929dab19f54564286b0dd42cc80b8aa37d28ea55436d6eba9233bdeb" Oct 07 14:13:28 crc kubenswrapper[4959]: I1007 14:13:28.387984 4959 scope.go:117] "RemoveContainer" containerID="612295e06c3281aebb30eac6545902cd9ab7558de4fad84ee6b9eb0cbdc8d549" Oct 07 14:13:29 crc kubenswrapper[4959]: I1007 14:13:29.061536 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-45k2h"] Oct 07 14:13:29 crc kubenswrapper[4959]: I1007 14:13:29.072686 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-45k2h"] Oct 07 14:13:30 crc kubenswrapper[4959]: I1007 14:13:30.033138 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-ff2nz"] Oct 07 14:13:30 crc kubenswrapper[4959]: I1007 14:13:30.043683 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-ff2nz"] Oct 07 14:13:30 crc kubenswrapper[4959]: I1007 14:13:30.338980 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc" event={"ID":"4d310c8e-5f3a-484f-ab93-ee382d288df8","Type":"ContainerStarted","Data":"f2f5e667a2b6d76a4b24d950849f384c46ece4c029f8e6fdd40a681134a2bc00"} Oct 07 14:13:30 crc kubenswrapper[4959]: I1007 14:13:30.359732 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc" podStartSLOduration=1.74690027 podStartE2EDuration="3.359716106s" podCreationTimestamp="2025-10-07 14:13:27 +0000 UTC" firstStartedPulling="2025-10-07 14:13:28.23235312 +0000 UTC m=+1690.315757485" lastFinishedPulling="2025-10-07 14:13:29.845168986 +0000 UTC m=+1691.928573321" observedRunningTime="2025-10-07 14:13:30.357760717 +0000 UTC m=+1692.441165042" watchObservedRunningTime="2025-10-07 14:13:30.359716106 +0000 UTC m=+1692.443120431" Oct 07 14:13:30 crc kubenswrapper[4959]: I1007 14:13:30.664117 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fcce932-e96f-4a33-9511-d84540bd98fa" path="/var/lib/kubelet/pods/2fcce932-e96f-4a33-9511-d84540bd98fa/volumes" Oct 07 14:13:30 crc kubenswrapper[4959]: I1007 14:13:30.665064 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a359776a-07ed-46e8-b20b-f7b7addaed8d" path="/var/lib/kubelet/pods/a359776a-07ed-46e8-b20b-f7b7addaed8d/volumes" Oct 07 14:13:32 crc kubenswrapper[4959]: I1007 14:13:32.033217 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-2hhl9"] Oct 07 14:13:32 crc kubenswrapper[4959]: I1007 14:13:32.039895 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-2hhl9"] Oct 07 14:13:32 crc kubenswrapper[4959]: I1007 14:13:32.667943 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3b8b961-3c0e-47dc-b28f-d55c9821ab34" path="/var/lib/kubelet/pods/c3b8b961-3c0e-47dc-b28f-d55c9821ab34/volumes" Oct 07 14:13:34 crc kubenswrapper[4959]: I1007 14:13:34.656867 4959 scope.go:117] "RemoveContainer" containerID="8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1" Oct 07 14:13:34 crc kubenswrapper[4959]: E1007 14:13:34.658242 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:13:37 crc kubenswrapper[4959]: I1007 14:13:37.028332 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-jfrdg"] Oct 07 14:13:37 crc kubenswrapper[4959]: I1007 14:13:37.037981 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-jfrdg"] Oct 07 14:13:38 crc kubenswrapper[4959]: I1007 14:13:38.028435 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-4sfdr"] Oct 07 14:13:38 crc kubenswrapper[4959]: I1007 14:13:38.053020 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-4sfdr"] Oct 07 14:13:38 crc kubenswrapper[4959]: I1007 14:13:38.679651 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b70a321-ffb7-429f-b825-4cd872ded9ff" path="/var/lib/kubelet/pods/7b70a321-ffb7-429f-b825-4cd872ded9ff/volumes" Oct 07 14:13:38 crc kubenswrapper[4959]: I1007 14:13:38.683544 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="806fce72-373c-4c34-8293-3051691eb55f" path="/var/lib/kubelet/pods/806fce72-373c-4c34-8293-3051691eb55f/volumes" Oct 07 14:13:45 crc kubenswrapper[4959]: I1007 14:13:45.654482 4959 scope.go:117] "RemoveContainer" containerID="8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1" Oct 07 14:13:45 crc kubenswrapper[4959]: E1007 14:13:45.654941 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:14:00 crc kubenswrapper[4959]: I1007 14:14:00.653759 4959 scope.go:117] "RemoveContainer" containerID="8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1" Oct 07 14:14:00 crc kubenswrapper[4959]: E1007 14:14:00.655268 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:14:11 crc kubenswrapper[4959]: I1007 14:14:11.654009 4959 scope.go:117] "RemoveContainer" containerID="8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1" Oct 07 14:14:11 crc kubenswrapper[4959]: E1007 14:14:11.654786 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:14:16 crc kubenswrapper[4959]: I1007 14:14:16.042794 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-jmxbs"] Oct 07 14:14:16 crc kubenswrapper[4959]: I1007 14:14:16.055348 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-77f42"] Oct 07 14:14:16 crc kubenswrapper[4959]: I1007 14:14:16.062152 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-g494x"] Oct 07 14:14:16 crc kubenswrapper[4959]: I1007 14:14:16.067884 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-g494x"] Oct 07 14:14:16 crc kubenswrapper[4959]: I1007 14:14:16.073891 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-jmxbs"] Oct 07 14:14:16 crc kubenswrapper[4959]: I1007 14:14:16.079003 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-77f42"] Oct 07 14:14:16 crc kubenswrapper[4959]: I1007 14:14:16.671474 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ab681da-c60c-415b-985c-ca36c4cbead0" path="/var/lib/kubelet/pods/7ab681da-c60c-415b-985c-ca36c4cbead0/volumes" Oct 07 14:14:16 crc kubenswrapper[4959]: I1007 14:14:16.672731 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="861069fb-97b5-4ae8-a7da-0f0c51ff50ac" path="/var/lib/kubelet/pods/861069fb-97b5-4ae8-a7da-0f0c51ff50ac/volumes" Oct 07 14:14:16 crc kubenswrapper[4959]: I1007 14:14:16.673790 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb1f3a07-1ea0-4d4d-aba7-bc1912e3a47c" path="/var/lib/kubelet/pods/bb1f3a07-1ea0-4d4d-aba7-bc1912e3a47c/volumes" Oct 07 14:14:22 crc kubenswrapper[4959]: I1007 14:14:22.653512 4959 scope.go:117] "RemoveContainer" containerID="8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1" Oct 07 14:14:22 crc kubenswrapper[4959]: E1007 14:14:22.654401 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:14:23 crc kubenswrapper[4959]: I1007 14:14:23.026835 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-3e47-account-create-h55ck"] Oct 07 14:14:23 crc kubenswrapper[4959]: I1007 14:14:23.033327 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-3e47-account-create-h55ck"] Oct 07 14:14:24 crc kubenswrapper[4959]: I1007 14:14:24.674826 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9743ad9-caa9-4bc4-97ef-fa52706ca1ae" path="/var/lib/kubelet/pods/c9743ad9-caa9-4bc4-97ef-fa52706ca1ae/volumes" Oct 07 14:14:28 crc kubenswrapper[4959]: I1007 14:14:28.512013 4959 scope.go:117] "RemoveContainer" containerID="8516f5e1fc246fe52543f3db0e6133ecf48e28d79a57e9be359176aea738d69f" Oct 07 14:14:28 crc kubenswrapper[4959]: I1007 14:14:28.551542 4959 scope.go:117] "RemoveContainer" containerID="6361a07ae0f69ab1a40e597bf3eb451fd4fbeedc2ddf0353d49a648fca005ca9" Oct 07 14:14:28 crc kubenswrapper[4959]: I1007 14:14:28.595907 4959 scope.go:117] "RemoveContainer" containerID="66798f105e40cb45703200b22d95be4846111976ce149d676b6c3f12a94de0f4" Oct 07 14:14:28 crc kubenswrapper[4959]: I1007 14:14:28.629373 4959 scope.go:117] "RemoveContainer" containerID="a838210ef105ae5f335860c3556e14b59e63b90c800bb0fc31efab2ec05db490" Oct 07 14:14:28 crc kubenswrapper[4959]: I1007 14:14:28.661340 4959 scope.go:117] "RemoveContainer" containerID="88d73562583ddda94182356c56de7337fd18fc33948a51321d001c62cd5d8ebc" Oct 07 14:14:28 crc kubenswrapper[4959]: I1007 14:14:28.719048 4959 scope.go:117] "RemoveContainer" containerID="59fac583237ac7d0d6ba7ead48a081c6075c175f02ac7bbc456ae75dabbac082" Oct 07 14:14:28 crc kubenswrapper[4959]: I1007 14:14:28.755296 4959 scope.go:117] "RemoveContainer" containerID="1c98cd878dfdc18d00933f385480ff055b149f850fe48ee3960f16c3ce508cb7" Oct 07 14:14:28 crc kubenswrapper[4959]: I1007 14:14:28.791800 4959 scope.go:117] "RemoveContainer" containerID="23adf200733196bc33a98c2ba5078225882e47f27d4fad79ecbc95256ca18655" Oct 07 14:14:28 crc kubenswrapper[4959]: I1007 14:14:28.809761 4959 scope.go:117] "RemoveContainer" containerID="d0594f269876109a127823477b2fc4b883d1d61793068015b09be8b59b07c22a" Oct 07 14:14:28 crc kubenswrapper[4959]: I1007 14:14:28.869233 4959 generic.go:334] "Generic (PLEG): container finished" podID="4d310c8e-5f3a-484f-ab93-ee382d288df8" containerID="f2f5e667a2b6d76a4b24d950849f384c46ece4c029f8e6fdd40a681134a2bc00" exitCode=2 Oct 07 14:14:28 crc kubenswrapper[4959]: I1007 14:14:28.869297 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc" event={"ID":"4d310c8e-5f3a-484f-ab93-ee382d288df8","Type":"ContainerDied","Data":"f2f5e667a2b6d76a4b24d950849f384c46ece4c029f8e6fdd40a681134a2bc00"} Oct 07 14:14:30 crc kubenswrapper[4959]: I1007 14:14:30.306320 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc" Oct 07 14:14:30 crc kubenswrapper[4959]: I1007 14:14:30.484316 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d310c8e-5f3a-484f-ab93-ee382d288df8-ssh-key\") pod \"4d310c8e-5f3a-484f-ab93-ee382d288df8\" (UID: \"4d310c8e-5f3a-484f-ab93-ee382d288df8\") " Oct 07 14:14:30 crc kubenswrapper[4959]: I1007 14:14:30.484607 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8gmf2\" (UniqueName: \"kubernetes.io/projected/4d310c8e-5f3a-484f-ab93-ee382d288df8-kube-api-access-8gmf2\") pod \"4d310c8e-5f3a-484f-ab93-ee382d288df8\" (UID: \"4d310c8e-5f3a-484f-ab93-ee382d288df8\") " Oct 07 14:14:30 crc kubenswrapper[4959]: I1007 14:14:30.484636 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d310c8e-5f3a-484f-ab93-ee382d288df8-inventory\") pod \"4d310c8e-5f3a-484f-ab93-ee382d288df8\" (UID: \"4d310c8e-5f3a-484f-ab93-ee382d288df8\") " Oct 07 14:14:30 crc kubenswrapper[4959]: I1007 14:14:30.496441 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d310c8e-5f3a-484f-ab93-ee382d288df8-kube-api-access-8gmf2" (OuterVolumeSpecName: "kube-api-access-8gmf2") pod "4d310c8e-5f3a-484f-ab93-ee382d288df8" (UID: "4d310c8e-5f3a-484f-ab93-ee382d288df8"). InnerVolumeSpecName "kube-api-access-8gmf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:14:30 crc kubenswrapper[4959]: I1007 14:14:30.510941 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d310c8e-5f3a-484f-ab93-ee382d288df8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4d310c8e-5f3a-484f-ab93-ee382d288df8" (UID: "4d310c8e-5f3a-484f-ab93-ee382d288df8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:14:30 crc kubenswrapper[4959]: I1007 14:14:30.538958 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d310c8e-5f3a-484f-ab93-ee382d288df8-inventory" (OuterVolumeSpecName: "inventory") pod "4d310c8e-5f3a-484f-ab93-ee382d288df8" (UID: "4d310c8e-5f3a-484f-ab93-ee382d288df8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:14:30 crc kubenswrapper[4959]: I1007 14:14:30.587567 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d310c8e-5f3a-484f-ab93-ee382d288df8-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 14:14:30 crc kubenswrapper[4959]: I1007 14:14:30.587621 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8gmf2\" (UniqueName: \"kubernetes.io/projected/4d310c8e-5f3a-484f-ab93-ee382d288df8-kube-api-access-8gmf2\") on node \"crc\" DevicePath \"\"" Oct 07 14:14:30 crc kubenswrapper[4959]: I1007 14:14:30.587646 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d310c8e-5f3a-484f-ab93-ee382d288df8-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 14:14:30 crc kubenswrapper[4959]: I1007 14:14:30.921691 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc" event={"ID":"4d310c8e-5f3a-484f-ab93-ee382d288df8","Type":"ContainerDied","Data":"938d12d11a6f1ae863bd84d588c222cd85cca4965d8283822d12e183b52d9dd2"} Oct 07 14:14:30 crc kubenswrapper[4959]: I1007 14:14:30.921764 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="938d12d11a6f1ae863bd84d588c222cd85cca4965d8283822d12e183b52d9dd2" Oct 07 14:14:30 crc kubenswrapper[4959]: I1007 14:14:30.921856 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc" Oct 07 14:14:33 crc kubenswrapper[4959]: I1007 14:14:33.653742 4959 scope.go:117] "RemoveContainer" containerID="8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1" Oct 07 14:14:33 crc kubenswrapper[4959]: E1007 14:14:33.654377 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:14:34 crc kubenswrapper[4959]: I1007 14:14:34.031150 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-1479-account-create-n9c4j"] Oct 07 14:14:34 crc kubenswrapper[4959]: I1007 14:14:34.038799 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-79f7-account-create-7n7bb"] Oct 07 14:14:34 crc kubenswrapper[4959]: I1007 14:14:34.046482 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-1479-account-create-n9c4j"] Oct 07 14:14:34 crc kubenswrapper[4959]: I1007 14:14:34.053364 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-79f7-account-create-7n7bb"] Oct 07 14:14:34 crc kubenswrapper[4959]: I1007 14:14:34.669792 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63e77afb-d7b9-46a0-8f3c-e12d8157812d" path="/var/lib/kubelet/pods/63e77afb-d7b9-46a0-8f3c-e12d8157812d/volumes" Oct 07 14:14:34 crc kubenswrapper[4959]: I1007 14:14:34.671216 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80610e88-4157-4a37-bbe6-97d422d5f97e" path="/var/lib/kubelet/pods/80610e88-4157-4a37-bbe6-97d422d5f97e/volumes" Oct 07 14:14:38 crc kubenswrapper[4959]: I1007 14:14:38.030539 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-km6rz"] Oct 07 14:14:38 crc kubenswrapper[4959]: E1007 14:14:38.031352 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d310c8e-5f3a-484f-ab93-ee382d288df8" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 07 14:14:38 crc kubenswrapper[4959]: I1007 14:14:38.031370 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d310c8e-5f3a-484f-ab93-ee382d288df8" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 07 14:14:38 crc kubenswrapper[4959]: I1007 14:14:38.031598 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d310c8e-5f3a-484f-ab93-ee382d288df8" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 07 14:14:38 crc kubenswrapper[4959]: I1007 14:14:38.032276 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-km6rz" Oct 07 14:14:38 crc kubenswrapper[4959]: I1007 14:14:38.034302 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 14:14:38 crc kubenswrapper[4959]: I1007 14:14:38.035762 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 14:14:38 crc kubenswrapper[4959]: I1007 14:14:38.036292 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-c6vmh" Oct 07 14:14:38 crc kubenswrapper[4959]: I1007 14:14:38.036401 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 14:14:38 crc kubenswrapper[4959]: I1007 14:14:38.053661 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-km6rz"] Oct 07 14:14:38 crc kubenswrapper[4959]: I1007 14:14:38.133583 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2wch\" (UniqueName: \"kubernetes.io/projected/ac13d4c5-d59f-4a95-9253-8417785bc949-kube-api-access-z2wch\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-km6rz\" (UID: \"ac13d4c5-d59f-4a95-9253-8417785bc949\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-km6rz" Oct 07 14:14:38 crc kubenswrapper[4959]: I1007 14:14:38.133784 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac13d4c5-d59f-4a95-9253-8417785bc949-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-km6rz\" (UID: \"ac13d4c5-d59f-4a95-9253-8417785bc949\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-km6rz" Oct 07 14:14:38 crc kubenswrapper[4959]: I1007 14:14:38.133832 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac13d4c5-d59f-4a95-9253-8417785bc949-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-km6rz\" (UID: \"ac13d4c5-d59f-4a95-9253-8417785bc949\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-km6rz" Oct 07 14:14:38 crc kubenswrapper[4959]: I1007 14:14:38.235640 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac13d4c5-d59f-4a95-9253-8417785bc949-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-km6rz\" (UID: \"ac13d4c5-d59f-4a95-9253-8417785bc949\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-km6rz" Oct 07 14:14:38 crc kubenswrapper[4959]: I1007 14:14:38.235692 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac13d4c5-d59f-4a95-9253-8417785bc949-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-km6rz\" (UID: \"ac13d4c5-d59f-4a95-9253-8417785bc949\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-km6rz" Oct 07 14:14:38 crc kubenswrapper[4959]: I1007 14:14:38.235830 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2wch\" (UniqueName: \"kubernetes.io/projected/ac13d4c5-d59f-4a95-9253-8417785bc949-kube-api-access-z2wch\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-km6rz\" (UID: \"ac13d4c5-d59f-4a95-9253-8417785bc949\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-km6rz" Oct 07 14:14:38 crc kubenswrapper[4959]: I1007 14:14:38.241367 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac13d4c5-d59f-4a95-9253-8417785bc949-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-km6rz\" (UID: \"ac13d4c5-d59f-4a95-9253-8417785bc949\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-km6rz" Oct 07 14:14:38 crc kubenswrapper[4959]: I1007 14:14:38.243228 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac13d4c5-d59f-4a95-9253-8417785bc949-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-km6rz\" (UID: \"ac13d4c5-d59f-4a95-9253-8417785bc949\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-km6rz" Oct 07 14:14:38 crc kubenswrapper[4959]: I1007 14:14:38.260910 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2wch\" (UniqueName: \"kubernetes.io/projected/ac13d4c5-d59f-4a95-9253-8417785bc949-kube-api-access-z2wch\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-km6rz\" (UID: \"ac13d4c5-d59f-4a95-9253-8417785bc949\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-km6rz" Oct 07 14:14:38 crc kubenswrapper[4959]: I1007 14:14:38.360336 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-km6rz" Oct 07 14:14:38 crc kubenswrapper[4959]: I1007 14:14:38.897069 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-km6rz"] Oct 07 14:14:38 crc kubenswrapper[4959]: I1007 14:14:38.988289 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-km6rz" event={"ID":"ac13d4c5-d59f-4a95-9253-8417785bc949","Type":"ContainerStarted","Data":"13c4ff031476723f9b4e98d4da1b8a8575ca1453ff9b2eb750659aae53371403"} Oct 07 14:14:41 crc kubenswrapper[4959]: I1007 14:14:41.004090 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-km6rz" event={"ID":"ac13d4c5-d59f-4a95-9253-8417785bc949","Type":"ContainerStarted","Data":"6540e28dc68f53ae9ebcd09d0b1dc6198da7c5ed82f03c4807c4af7669f43176"} Oct 07 14:14:41 crc kubenswrapper[4959]: I1007 14:14:41.022881 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-km6rz" podStartSLOduration=2.06384513 podStartE2EDuration="3.02286108s" podCreationTimestamp="2025-10-07 14:14:38 +0000 UTC" firstStartedPulling="2025-10-07 14:14:38.901134786 +0000 UTC m=+1760.984539111" lastFinishedPulling="2025-10-07 14:14:39.860150736 +0000 UTC m=+1761.943555061" observedRunningTime="2025-10-07 14:14:41.016561142 +0000 UTC m=+1763.099965467" watchObservedRunningTime="2025-10-07 14:14:41.02286108 +0000 UTC m=+1763.106265405" Oct 07 14:14:47 crc kubenswrapper[4959]: I1007 14:14:47.654299 4959 scope.go:117] "RemoveContainer" containerID="8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1" Oct 07 14:14:47 crc kubenswrapper[4959]: E1007 14:14:47.655251 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:14:59 crc kubenswrapper[4959]: I1007 14:14:59.653270 4959 scope.go:117] "RemoveContainer" containerID="8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1" Oct 07 14:14:59 crc kubenswrapper[4959]: E1007 14:14:59.654010 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:15:00 crc kubenswrapper[4959]: I1007 14:15:00.158679 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330775-flfr7"] Oct 07 14:15:00 crc kubenswrapper[4959]: I1007 14:15:00.164299 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330775-flfr7" Oct 07 14:15:00 crc kubenswrapper[4959]: I1007 14:15:00.168609 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 07 14:15:00 crc kubenswrapper[4959]: I1007 14:15:00.168841 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 07 14:15:00 crc kubenswrapper[4959]: I1007 14:15:00.172962 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330775-flfr7"] Oct 07 14:15:00 crc kubenswrapper[4959]: I1007 14:15:00.267701 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ba6d1d1d-89bf-4c7c-9266-cebc0505e53c-secret-volume\") pod \"collect-profiles-29330775-flfr7\" (UID: \"ba6d1d1d-89bf-4c7c-9266-cebc0505e53c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330775-flfr7" Oct 07 14:15:00 crc kubenswrapper[4959]: I1007 14:15:00.267770 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sms6\" (UniqueName: \"kubernetes.io/projected/ba6d1d1d-89bf-4c7c-9266-cebc0505e53c-kube-api-access-8sms6\") pod \"collect-profiles-29330775-flfr7\" (UID: \"ba6d1d1d-89bf-4c7c-9266-cebc0505e53c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330775-flfr7" Oct 07 14:15:00 crc kubenswrapper[4959]: I1007 14:15:00.268025 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ba6d1d1d-89bf-4c7c-9266-cebc0505e53c-config-volume\") pod \"collect-profiles-29330775-flfr7\" (UID: \"ba6d1d1d-89bf-4c7c-9266-cebc0505e53c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330775-flfr7" Oct 07 14:15:00 crc kubenswrapper[4959]: I1007 14:15:00.370050 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ba6d1d1d-89bf-4c7c-9266-cebc0505e53c-config-volume\") pod \"collect-profiles-29330775-flfr7\" (UID: \"ba6d1d1d-89bf-4c7c-9266-cebc0505e53c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330775-flfr7" Oct 07 14:15:00 crc kubenswrapper[4959]: I1007 14:15:00.370726 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ba6d1d1d-89bf-4c7c-9266-cebc0505e53c-secret-volume\") pod \"collect-profiles-29330775-flfr7\" (UID: \"ba6d1d1d-89bf-4c7c-9266-cebc0505e53c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330775-flfr7" Oct 07 14:15:00 crc kubenswrapper[4959]: I1007 14:15:00.370786 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8sms6\" (UniqueName: \"kubernetes.io/projected/ba6d1d1d-89bf-4c7c-9266-cebc0505e53c-kube-api-access-8sms6\") pod \"collect-profiles-29330775-flfr7\" (UID: \"ba6d1d1d-89bf-4c7c-9266-cebc0505e53c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330775-flfr7" Oct 07 14:15:00 crc kubenswrapper[4959]: I1007 14:15:00.371321 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ba6d1d1d-89bf-4c7c-9266-cebc0505e53c-config-volume\") pod \"collect-profiles-29330775-flfr7\" (UID: \"ba6d1d1d-89bf-4c7c-9266-cebc0505e53c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330775-flfr7" Oct 07 14:15:00 crc kubenswrapper[4959]: I1007 14:15:00.383921 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ba6d1d1d-89bf-4c7c-9266-cebc0505e53c-secret-volume\") pod \"collect-profiles-29330775-flfr7\" (UID: \"ba6d1d1d-89bf-4c7c-9266-cebc0505e53c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330775-flfr7" Oct 07 14:15:00 crc kubenswrapper[4959]: I1007 14:15:00.386393 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8sms6\" (UniqueName: \"kubernetes.io/projected/ba6d1d1d-89bf-4c7c-9266-cebc0505e53c-kube-api-access-8sms6\") pod \"collect-profiles-29330775-flfr7\" (UID: \"ba6d1d1d-89bf-4c7c-9266-cebc0505e53c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330775-flfr7" Oct 07 14:15:00 crc kubenswrapper[4959]: I1007 14:15:00.492154 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330775-flfr7" Oct 07 14:15:01 crc kubenswrapper[4959]: I1007 14:15:01.008841 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330775-flfr7"] Oct 07 14:15:01 crc kubenswrapper[4959]: I1007 14:15:01.171353 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330775-flfr7" event={"ID":"ba6d1d1d-89bf-4c7c-9266-cebc0505e53c","Type":"ContainerStarted","Data":"b6f4ec8da839a2822fb6b7a623270f34514e24e92505ed1b7aae978e10aea363"} Oct 07 14:15:02 crc kubenswrapper[4959]: I1007 14:15:02.182547 4959 generic.go:334] "Generic (PLEG): container finished" podID="ba6d1d1d-89bf-4c7c-9266-cebc0505e53c" containerID="439abfcf60e7afaf0f1b838e06043c104500921fe892d2fc7c1013dd14ad616d" exitCode=0 Oct 07 14:15:02 crc kubenswrapper[4959]: I1007 14:15:02.182600 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330775-flfr7" event={"ID":"ba6d1d1d-89bf-4c7c-9266-cebc0505e53c","Type":"ContainerDied","Data":"439abfcf60e7afaf0f1b838e06043c104500921fe892d2fc7c1013dd14ad616d"} Oct 07 14:15:03 crc kubenswrapper[4959]: I1007 14:15:03.505711 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330775-flfr7" Oct 07 14:15:03 crc kubenswrapper[4959]: I1007 14:15:03.632958 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ba6d1d1d-89bf-4c7c-9266-cebc0505e53c-secret-volume\") pod \"ba6d1d1d-89bf-4c7c-9266-cebc0505e53c\" (UID: \"ba6d1d1d-89bf-4c7c-9266-cebc0505e53c\") " Oct 07 14:15:03 crc kubenswrapper[4959]: I1007 14:15:03.633065 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8sms6\" (UniqueName: \"kubernetes.io/projected/ba6d1d1d-89bf-4c7c-9266-cebc0505e53c-kube-api-access-8sms6\") pod \"ba6d1d1d-89bf-4c7c-9266-cebc0505e53c\" (UID: \"ba6d1d1d-89bf-4c7c-9266-cebc0505e53c\") " Oct 07 14:15:03 crc kubenswrapper[4959]: I1007 14:15:03.633219 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ba6d1d1d-89bf-4c7c-9266-cebc0505e53c-config-volume\") pod \"ba6d1d1d-89bf-4c7c-9266-cebc0505e53c\" (UID: \"ba6d1d1d-89bf-4c7c-9266-cebc0505e53c\") " Oct 07 14:15:03 crc kubenswrapper[4959]: I1007 14:15:03.634147 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba6d1d1d-89bf-4c7c-9266-cebc0505e53c-config-volume" (OuterVolumeSpecName: "config-volume") pod "ba6d1d1d-89bf-4c7c-9266-cebc0505e53c" (UID: "ba6d1d1d-89bf-4c7c-9266-cebc0505e53c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:15:03 crc kubenswrapper[4959]: I1007 14:15:03.639295 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba6d1d1d-89bf-4c7c-9266-cebc0505e53c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ba6d1d1d-89bf-4c7c-9266-cebc0505e53c" (UID: "ba6d1d1d-89bf-4c7c-9266-cebc0505e53c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:15:03 crc kubenswrapper[4959]: I1007 14:15:03.640060 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba6d1d1d-89bf-4c7c-9266-cebc0505e53c-kube-api-access-8sms6" (OuterVolumeSpecName: "kube-api-access-8sms6") pod "ba6d1d1d-89bf-4c7c-9266-cebc0505e53c" (UID: "ba6d1d1d-89bf-4c7c-9266-cebc0505e53c"). InnerVolumeSpecName "kube-api-access-8sms6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:15:03 crc kubenswrapper[4959]: I1007 14:15:03.735734 4959 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ba6d1d1d-89bf-4c7c-9266-cebc0505e53c-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 07 14:15:03 crc kubenswrapper[4959]: I1007 14:15:03.735777 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8sms6\" (UniqueName: \"kubernetes.io/projected/ba6d1d1d-89bf-4c7c-9266-cebc0505e53c-kube-api-access-8sms6\") on node \"crc\" DevicePath \"\"" Oct 07 14:15:03 crc kubenswrapper[4959]: I1007 14:15:03.735793 4959 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ba6d1d1d-89bf-4c7c-9266-cebc0505e53c-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 14:15:04 crc kubenswrapper[4959]: I1007 14:15:04.209719 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330775-flfr7" event={"ID":"ba6d1d1d-89bf-4c7c-9266-cebc0505e53c","Type":"ContainerDied","Data":"b6f4ec8da839a2822fb6b7a623270f34514e24e92505ed1b7aae978e10aea363"} Oct 07 14:15:04 crc kubenswrapper[4959]: I1007 14:15:04.209766 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6f4ec8da839a2822fb6b7a623270f34514e24e92505ed1b7aae978e10aea363" Oct 07 14:15:04 crc kubenswrapper[4959]: I1007 14:15:04.209868 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330775-flfr7" Oct 07 14:15:07 crc kubenswrapper[4959]: I1007 14:15:07.039001 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6djrf"] Oct 07 14:15:07 crc kubenswrapper[4959]: I1007 14:15:07.050263 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6djrf"] Oct 07 14:15:08 crc kubenswrapper[4959]: I1007 14:15:08.664364 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bfaeeb0-e041-452a-aa79-9a65f2b1def8" path="/var/lib/kubelet/pods/7bfaeeb0-e041-452a-aa79-9a65f2b1def8/volumes" Oct 07 14:15:10 crc kubenswrapper[4959]: I1007 14:15:10.653761 4959 scope.go:117] "RemoveContainer" containerID="8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1" Oct 07 14:15:10 crc kubenswrapper[4959]: E1007 14:15:10.655047 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:15:24 crc kubenswrapper[4959]: I1007 14:15:24.653904 4959 scope.go:117] "RemoveContainer" containerID="8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1" Oct 07 14:15:24 crc kubenswrapper[4959]: E1007 14:15:24.655037 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:15:25 crc kubenswrapper[4959]: I1007 14:15:25.377475 4959 generic.go:334] "Generic (PLEG): container finished" podID="ac13d4c5-d59f-4a95-9253-8417785bc949" containerID="6540e28dc68f53ae9ebcd09d0b1dc6198da7c5ed82f03c4807c4af7669f43176" exitCode=0 Oct 07 14:15:25 crc kubenswrapper[4959]: I1007 14:15:25.377565 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-km6rz" event={"ID":"ac13d4c5-d59f-4a95-9253-8417785bc949","Type":"ContainerDied","Data":"6540e28dc68f53ae9ebcd09d0b1dc6198da7c5ed82f03c4807c4af7669f43176"} Oct 07 14:15:26 crc kubenswrapper[4959]: I1007 14:15:26.765310 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-km6rz" Oct 07 14:15:26 crc kubenswrapper[4959]: I1007 14:15:26.885351 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac13d4c5-d59f-4a95-9253-8417785bc949-ssh-key\") pod \"ac13d4c5-d59f-4a95-9253-8417785bc949\" (UID: \"ac13d4c5-d59f-4a95-9253-8417785bc949\") " Oct 07 14:15:26 crc kubenswrapper[4959]: I1007 14:15:26.885544 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac13d4c5-d59f-4a95-9253-8417785bc949-inventory\") pod \"ac13d4c5-d59f-4a95-9253-8417785bc949\" (UID: \"ac13d4c5-d59f-4a95-9253-8417785bc949\") " Oct 07 14:15:26 crc kubenswrapper[4959]: I1007 14:15:26.885584 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2wch\" (UniqueName: \"kubernetes.io/projected/ac13d4c5-d59f-4a95-9253-8417785bc949-kube-api-access-z2wch\") pod \"ac13d4c5-d59f-4a95-9253-8417785bc949\" (UID: \"ac13d4c5-d59f-4a95-9253-8417785bc949\") " Oct 07 14:15:26 crc kubenswrapper[4959]: I1007 14:15:26.891739 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac13d4c5-d59f-4a95-9253-8417785bc949-kube-api-access-z2wch" (OuterVolumeSpecName: "kube-api-access-z2wch") pod "ac13d4c5-d59f-4a95-9253-8417785bc949" (UID: "ac13d4c5-d59f-4a95-9253-8417785bc949"). InnerVolumeSpecName "kube-api-access-z2wch". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:15:26 crc kubenswrapper[4959]: I1007 14:15:26.911291 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac13d4c5-d59f-4a95-9253-8417785bc949-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ac13d4c5-d59f-4a95-9253-8417785bc949" (UID: "ac13d4c5-d59f-4a95-9253-8417785bc949"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:15:26 crc kubenswrapper[4959]: I1007 14:15:26.921428 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac13d4c5-d59f-4a95-9253-8417785bc949-inventory" (OuterVolumeSpecName: "inventory") pod "ac13d4c5-d59f-4a95-9253-8417785bc949" (UID: "ac13d4c5-d59f-4a95-9253-8417785bc949"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:15:26 crc kubenswrapper[4959]: I1007 14:15:26.988570 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac13d4c5-d59f-4a95-9253-8417785bc949-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 14:15:26 crc kubenswrapper[4959]: I1007 14:15:26.988612 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac13d4c5-d59f-4a95-9253-8417785bc949-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 14:15:26 crc kubenswrapper[4959]: I1007 14:15:26.988626 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2wch\" (UniqueName: \"kubernetes.io/projected/ac13d4c5-d59f-4a95-9253-8417785bc949-kube-api-access-z2wch\") on node \"crc\" DevicePath \"\"" Oct 07 14:15:27 crc kubenswrapper[4959]: I1007 14:15:27.393167 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-km6rz" event={"ID":"ac13d4c5-d59f-4a95-9253-8417785bc949","Type":"ContainerDied","Data":"13c4ff031476723f9b4e98d4da1b8a8575ca1453ff9b2eb750659aae53371403"} Oct 07 14:15:27 crc kubenswrapper[4959]: I1007 14:15:27.393517 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13c4ff031476723f9b4e98d4da1b8a8575ca1453ff9b2eb750659aae53371403" Oct 07 14:15:27 crc kubenswrapper[4959]: I1007 14:15:27.393231 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-km6rz" Oct 07 14:15:27 crc kubenswrapper[4959]: I1007 14:15:27.468616 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-h5brv"] Oct 07 14:15:27 crc kubenswrapper[4959]: E1007 14:15:27.469053 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac13d4c5-d59f-4a95-9253-8417785bc949" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 07 14:15:27 crc kubenswrapper[4959]: I1007 14:15:27.469075 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac13d4c5-d59f-4a95-9253-8417785bc949" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 07 14:15:27 crc kubenswrapper[4959]: E1007 14:15:27.469122 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba6d1d1d-89bf-4c7c-9266-cebc0505e53c" containerName="collect-profiles" Oct 07 14:15:27 crc kubenswrapper[4959]: I1007 14:15:27.469132 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba6d1d1d-89bf-4c7c-9266-cebc0505e53c" containerName="collect-profiles" Oct 07 14:15:27 crc kubenswrapper[4959]: I1007 14:15:27.469398 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba6d1d1d-89bf-4c7c-9266-cebc0505e53c" containerName="collect-profiles" Oct 07 14:15:27 crc kubenswrapper[4959]: I1007 14:15:27.469441 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac13d4c5-d59f-4a95-9253-8417785bc949" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 07 14:15:27 crc kubenswrapper[4959]: I1007 14:15:27.470134 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-h5brv" Oct 07 14:15:27 crc kubenswrapper[4959]: I1007 14:15:27.474932 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 14:15:27 crc kubenswrapper[4959]: I1007 14:15:27.475133 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 14:15:27 crc kubenswrapper[4959]: I1007 14:15:27.475243 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 14:15:27 crc kubenswrapper[4959]: I1007 14:15:27.475425 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-c6vmh" Oct 07 14:15:27 crc kubenswrapper[4959]: I1007 14:15:27.481014 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-h5brv"] Oct 07 14:15:27 crc kubenswrapper[4959]: I1007 14:15:27.500156 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/52b87677-a4da-4477-9a8f-8a5a656ce4bc-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-h5brv\" (UID: \"52b87677-a4da-4477-9a8f-8a5a656ce4bc\") " pod="openstack/ssh-known-hosts-edpm-deployment-h5brv" Oct 07 14:15:27 crc kubenswrapper[4959]: I1007 14:15:27.500213 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzqmd\" (UniqueName: \"kubernetes.io/projected/52b87677-a4da-4477-9a8f-8a5a656ce4bc-kube-api-access-kzqmd\") pod \"ssh-known-hosts-edpm-deployment-h5brv\" (UID: \"52b87677-a4da-4477-9a8f-8a5a656ce4bc\") " pod="openstack/ssh-known-hosts-edpm-deployment-h5brv" Oct 07 14:15:27 crc kubenswrapper[4959]: I1007 14:15:27.500644 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/52b87677-a4da-4477-9a8f-8a5a656ce4bc-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-h5brv\" (UID: \"52b87677-a4da-4477-9a8f-8a5a656ce4bc\") " pod="openstack/ssh-known-hosts-edpm-deployment-h5brv" Oct 07 14:15:27 crc kubenswrapper[4959]: I1007 14:15:27.601909 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/52b87677-a4da-4477-9a8f-8a5a656ce4bc-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-h5brv\" (UID: \"52b87677-a4da-4477-9a8f-8a5a656ce4bc\") " pod="openstack/ssh-known-hosts-edpm-deployment-h5brv" Oct 07 14:15:27 crc kubenswrapper[4959]: I1007 14:15:27.601982 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/52b87677-a4da-4477-9a8f-8a5a656ce4bc-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-h5brv\" (UID: \"52b87677-a4da-4477-9a8f-8a5a656ce4bc\") " pod="openstack/ssh-known-hosts-edpm-deployment-h5brv" Oct 07 14:15:27 crc kubenswrapper[4959]: I1007 14:15:27.602001 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzqmd\" (UniqueName: \"kubernetes.io/projected/52b87677-a4da-4477-9a8f-8a5a656ce4bc-kube-api-access-kzqmd\") pod \"ssh-known-hosts-edpm-deployment-h5brv\" (UID: \"52b87677-a4da-4477-9a8f-8a5a656ce4bc\") " pod="openstack/ssh-known-hosts-edpm-deployment-h5brv" Oct 07 14:15:27 crc kubenswrapper[4959]: I1007 14:15:27.607429 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/52b87677-a4da-4477-9a8f-8a5a656ce4bc-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-h5brv\" (UID: \"52b87677-a4da-4477-9a8f-8a5a656ce4bc\") " pod="openstack/ssh-known-hosts-edpm-deployment-h5brv" Oct 07 14:15:27 crc kubenswrapper[4959]: I1007 14:15:27.607449 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/52b87677-a4da-4477-9a8f-8a5a656ce4bc-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-h5brv\" (UID: \"52b87677-a4da-4477-9a8f-8a5a656ce4bc\") " pod="openstack/ssh-known-hosts-edpm-deployment-h5brv" Oct 07 14:15:27 crc kubenswrapper[4959]: I1007 14:15:27.620030 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzqmd\" (UniqueName: \"kubernetes.io/projected/52b87677-a4da-4477-9a8f-8a5a656ce4bc-kube-api-access-kzqmd\") pod \"ssh-known-hosts-edpm-deployment-h5brv\" (UID: \"52b87677-a4da-4477-9a8f-8a5a656ce4bc\") " pod="openstack/ssh-known-hosts-edpm-deployment-h5brv" Oct 07 14:15:27 crc kubenswrapper[4959]: I1007 14:15:27.797205 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-h5brv" Oct 07 14:15:28 crc kubenswrapper[4959]: I1007 14:15:28.312149 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-h5brv"] Oct 07 14:15:28 crc kubenswrapper[4959]: W1007 14:15:28.326335 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52b87677_a4da_4477_9a8f_8a5a656ce4bc.slice/crio-fe8239aa2dc7123bca23d7650572c7a69d7b2a4183c543c32b05522e72b3741e WatchSource:0}: Error finding container fe8239aa2dc7123bca23d7650572c7a69d7b2a4183c543c32b05522e72b3741e: Status 404 returned error can't find the container with id fe8239aa2dc7123bca23d7650572c7a69d7b2a4183c543c32b05522e72b3741e Oct 07 14:15:28 crc kubenswrapper[4959]: I1007 14:15:28.401010 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-h5brv" event={"ID":"52b87677-a4da-4477-9a8f-8a5a656ce4bc","Type":"ContainerStarted","Data":"fe8239aa2dc7123bca23d7650572c7a69d7b2a4183c543c32b05522e72b3741e"} Oct 07 14:15:29 crc kubenswrapper[4959]: I1007 14:15:29.001252 4959 scope.go:117] "RemoveContainer" containerID="5416daa5ada46ac3981c298e85ab9457898991c449ad80b24295028e6d00fe48" Oct 07 14:15:29 crc kubenswrapper[4959]: I1007 14:15:29.028473 4959 scope.go:117] "RemoveContainer" containerID="0d012cf781bb879e672e2a7388909f33c2f8516a32238cb1e064ce47451241ef" Oct 07 14:15:29 crc kubenswrapper[4959]: I1007 14:15:29.121849 4959 scope.go:117] "RemoveContainer" containerID="4ce2cc6d5f2da117ae8b950eb63cbd16aab79c2ee3381bed94410cb3b7149e32" Oct 07 14:15:30 crc kubenswrapper[4959]: I1007 14:15:30.039342 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-qltzt"] Oct 07 14:15:30 crc kubenswrapper[4959]: I1007 14:15:30.046907 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-qltzt"] Oct 07 14:15:30 crc kubenswrapper[4959]: I1007 14:15:30.663471 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f" path="/var/lib/kubelet/pods/1d6ad56e-d1a4-44c6-bc11-7f593e2afc1f/volumes" Oct 07 14:15:31 crc kubenswrapper[4959]: I1007 14:15:31.438077 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-h5brv" event={"ID":"52b87677-a4da-4477-9a8f-8a5a656ce4bc","Type":"ContainerStarted","Data":"18bd1574eb0bda3cf5dd8824be4eba6c0e4e2153c966f6cd54609346fb1a8a09"} Oct 07 14:15:31 crc kubenswrapper[4959]: I1007 14:15:31.460984 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-h5brv" podStartSLOduration=2.6531065050000002 podStartE2EDuration="4.460959878s" podCreationTimestamp="2025-10-07 14:15:27 +0000 UTC" firstStartedPulling="2025-10-07 14:15:28.328609209 +0000 UTC m=+1810.412013534" lastFinishedPulling="2025-10-07 14:15:30.136462582 +0000 UTC m=+1812.219866907" observedRunningTime="2025-10-07 14:15:31.456844365 +0000 UTC m=+1813.540248690" watchObservedRunningTime="2025-10-07 14:15:31.460959878 +0000 UTC m=+1813.544364203" Oct 07 14:15:34 crc kubenswrapper[4959]: I1007 14:15:34.026291 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-bkbj4"] Oct 07 14:15:34 crc kubenswrapper[4959]: I1007 14:15:34.033530 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-bkbj4"] Oct 07 14:15:34 crc kubenswrapper[4959]: I1007 14:15:34.664993 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="859a162a-5f11-416b-8156-bf39ba6bb53e" path="/var/lib/kubelet/pods/859a162a-5f11-416b-8156-bf39ba6bb53e/volumes" Oct 07 14:15:38 crc kubenswrapper[4959]: I1007 14:15:38.494074 4959 generic.go:334] "Generic (PLEG): container finished" podID="52b87677-a4da-4477-9a8f-8a5a656ce4bc" containerID="18bd1574eb0bda3cf5dd8824be4eba6c0e4e2153c966f6cd54609346fb1a8a09" exitCode=0 Oct 07 14:15:38 crc kubenswrapper[4959]: I1007 14:15:38.494184 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-h5brv" event={"ID":"52b87677-a4da-4477-9a8f-8a5a656ce4bc","Type":"ContainerDied","Data":"18bd1574eb0bda3cf5dd8824be4eba6c0e4e2153c966f6cd54609346fb1a8a09"} Oct 07 14:15:39 crc kubenswrapper[4959]: I1007 14:15:39.654864 4959 scope.go:117] "RemoveContainer" containerID="8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1" Oct 07 14:15:39 crc kubenswrapper[4959]: E1007 14:15:39.655477 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:15:39 crc kubenswrapper[4959]: I1007 14:15:39.929266 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-h5brv" Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.020558 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzqmd\" (UniqueName: \"kubernetes.io/projected/52b87677-a4da-4477-9a8f-8a5a656ce4bc-kube-api-access-kzqmd\") pod \"52b87677-a4da-4477-9a8f-8a5a656ce4bc\" (UID: \"52b87677-a4da-4477-9a8f-8a5a656ce4bc\") " Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.020617 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/52b87677-a4da-4477-9a8f-8a5a656ce4bc-inventory-0\") pod \"52b87677-a4da-4477-9a8f-8a5a656ce4bc\" (UID: \"52b87677-a4da-4477-9a8f-8a5a656ce4bc\") " Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.020724 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/52b87677-a4da-4477-9a8f-8a5a656ce4bc-ssh-key-openstack-edpm-ipam\") pod \"52b87677-a4da-4477-9a8f-8a5a656ce4bc\" (UID: \"52b87677-a4da-4477-9a8f-8a5a656ce4bc\") " Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.026539 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52b87677-a4da-4477-9a8f-8a5a656ce4bc-kube-api-access-kzqmd" (OuterVolumeSpecName: "kube-api-access-kzqmd") pod "52b87677-a4da-4477-9a8f-8a5a656ce4bc" (UID: "52b87677-a4da-4477-9a8f-8a5a656ce4bc"). InnerVolumeSpecName "kube-api-access-kzqmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.047637 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52b87677-a4da-4477-9a8f-8a5a656ce4bc-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "52b87677-a4da-4477-9a8f-8a5a656ce4bc" (UID: "52b87677-a4da-4477-9a8f-8a5a656ce4bc"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.051674 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52b87677-a4da-4477-9a8f-8a5a656ce4bc-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "52b87677-a4da-4477-9a8f-8a5a656ce4bc" (UID: "52b87677-a4da-4477-9a8f-8a5a656ce4bc"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.122349 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/52b87677-a4da-4477-9a8f-8a5a656ce4bc-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.122392 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzqmd\" (UniqueName: \"kubernetes.io/projected/52b87677-a4da-4477-9a8f-8a5a656ce4bc-kube-api-access-kzqmd\") on node \"crc\" DevicePath \"\"" Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.122407 4959 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/52b87677-a4da-4477-9a8f-8a5a656ce4bc-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.514531 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-h5brv" event={"ID":"52b87677-a4da-4477-9a8f-8a5a656ce4bc","Type":"ContainerDied","Data":"fe8239aa2dc7123bca23d7650572c7a69d7b2a4183c543c32b05522e72b3741e"} Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.514576 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-h5brv" Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.514585 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe8239aa2dc7123bca23d7650572c7a69d7b2a4183c543c32b05522e72b3741e" Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.591136 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkmr9"] Oct 07 14:15:40 crc kubenswrapper[4959]: E1007 14:15:40.591510 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52b87677-a4da-4477-9a8f-8a5a656ce4bc" containerName="ssh-known-hosts-edpm-deployment" Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.591528 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="52b87677-a4da-4477-9a8f-8a5a656ce4bc" containerName="ssh-known-hosts-edpm-deployment" Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.591703 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="52b87677-a4da-4477-9a8f-8a5a656ce4bc" containerName="ssh-known-hosts-edpm-deployment" Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.593745 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkmr9" Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.598929 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.601504 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.601572 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-c6vmh" Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.601672 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.605770 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkmr9"] Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.627616 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88e3b980-0df4-43e6-bdf1-cab57e2732dc-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jkmr9\" (UID: \"88e3b980-0df4-43e6-bdf1-cab57e2732dc\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkmr9" Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.627852 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxqvl\" (UniqueName: \"kubernetes.io/projected/88e3b980-0df4-43e6-bdf1-cab57e2732dc-kube-api-access-hxqvl\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jkmr9\" (UID: \"88e3b980-0df4-43e6-bdf1-cab57e2732dc\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkmr9" Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.627935 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88e3b980-0df4-43e6-bdf1-cab57e2732dc-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jkmr9\" (UID: \"88e3b980-0df4-43e6-bdf1-cab57e2732dc\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkmr9" Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.729511 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxqvl\" (UniqueName: \"kubernetes.io/projected/88e3b980-0df4-43e6-bdf1-cab57e2732dc-kube-api-access-hxqvl\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jkmr9\" (UID: \"88e3b980-0df4-43e6-bdf1-cab57e2732dc\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkmr9" Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.729603 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88e3b980-0df4-43e6-bdf1-cab57e2732dc-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jkmr9\" (UID: \"88e3b980-0df4-43e6-bdf1-cab57e2732dc\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkmr9" Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.729684 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88e3b980-0df4-43e6-bdf1-cab57e2732dc-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jkmr9\" (UID: \"88e3b980-0df4-43e6-bdf1-cab57e2732dc\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkmr9" Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.733017 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88e3b980-0df4-43e6-bdf1-cab57e2732dc-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jkmr9\" (UID: \"88e3b980-0df4-43e6-bdf1-cab57e2732dc\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkmr9" Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.733376 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88e3b980-0df4-43e6-bdf1-cab57e2732dc-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jkmr9\" (UID: \"88e3b980-0df4-43e6-bdf1-cab57e2732dc\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkmr9" Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.744786 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxqvl\" (UniqueName: \"kubernetes.io/projected/88e3b980-0df4-43e6-bdf1-cab57e2732dc-kube-api-access-hxqvl\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jkmr9\" (UID: \"88e3b980-0df4-43e6-bdf1-cab57e2732dc\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkmr9" Oct 07 14:15:40 crc kubenswrapper[4959]: I1007 14:15:40.910960 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkmr9" Oct 07 14:15:41 crc kubenswrapper[4959]: I1007 14:15:41.411374 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkmr9"] Oct 07 14:15:41 crc kubenswrapper[4959]: I1007 14:15:41.522508 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkmr9" event={"ID":"88e3b980-0df4-43e6-bdf1-cab57e2732dc","Type":"ContainerStarted","Data":"ba1e12932fb07bb4afafdab6a1993ac71c43498f2e76bf016f2096bba7c66303"} Oct 07 14:15:43 crc kubenswrapper[4959]: I1007 14:15:43.538117 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkmr9" event={"ID":"88e3b980-0df4-43e6-bdf1-cab57e2732dc","Type":"ContainerStarted","Data":"fdbec70c8246df00d7cf23865b4a11be9cf1cfbcfd29e732d8e4d539b6c8bb0e"} Oct 07 14:15:50 crc kubenswrapper[4959]: I1007 14:15:50.654236 4959 scope.go:117] "RemoveContainer" containerID="8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1" Oct 07 14:15:50 crc kubenswrapper[4959]: E1007 14:15:50.655033 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:15:52 crc kubenswrapper[4959]: I1007 14:15:52.609285 4959 generic.go:334] "Generic (PLEG): container finished" podID="88e3b980-0df4-43e6-bdf1-cab57e2732dc" containerID="fdbec70c8246df00d7cf23865b4a11be9cf1cfbcfd29e732d8e4d539b6c8bb0e" exitCode=0 Oct 07 14:15:52 crc kubenswrapper[4959]: I1007 14:15:52.609366 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkmr9" event={"ID":"88e3b980-0df4-43e6-bdf1-cab57e2732dc","Type":"ContainerDied","Data":"fdbec70c8246df00d7cf23865b4a11be9cf1cfbcfd29e732d8e4d539b6c8bb0e"} Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.030741 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkmr9" Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.080674 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxqvl\" (UniqueName: \"kubernetes.io/projected/88e3b980-0df4-43e6-bdf1-cab57e2732dc-kube-api-access-hxqvl\") pod \"88e3b980-0df4-43e6-bdf1-cab57e2732dc\" (UID: \"88e3b980-0df4-43e6-bdf1-cab57e2732dc\") " Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.080737 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88e3b980-0df4-43e6-bdf1-cab57e2732dc-ssh-key\") pod \"88e3b980-0df4-43e6-bdf1-cab57e2732dc\" (UID: \"88e3b980-0df4-43e6-bdf1-cab57e2732dc\") " Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.080807 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88e3b980-0df4-43e6-bdf1-cab57e2732dc-inventory\") pod \"88e3b980-0df4-43e6-bdf1-cab57e2732dc\" (UID: \"88e3b980-0df4-43e6-bdf1-cab57e2732dc\") " Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.085975 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88e3b980-0df4-43e6-bdf1-cab57e2732dc-kube-api-access-hxqvl" (OuterVolumeSpecName: "kube-api-access-hxqvl") pod "88e3b980-0df4-43e6-bdf1-cab57e2732dc" (UID: "88e3b980-0df4-43e6-bdf1-cab57e2732dc"). InnerVolumeSpecName "kube-api-access-hxqvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.107769 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88e3b980-0df4-43e6-bdf1-cab57e2732dc-inventory" (OuterVolumeSpecName: "inventory") pod "88e3b980-0df4-43e6-bdf1-cab57e2732dc" (UID: "88e3b980-0df4-43e6-bdf1-cab57e2732dc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.109146 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88e3b980-0df4-43e6-bdf1-cab57e2732dc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "88e3b980-0df4-43e6-bdf1-cab57e2732dc" (UID: "88e3b980-0df4-43e6-bdf1-cab57e2732dc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.182486 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxqvl\" (UniqueName: \"kubernetes.io/projected/88e3b980-0df4-43e6-bdf1-cab57e2732dc-kube-api-access-hxqvl\") on node \"crc\" DevicePath \"\"" Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.182519 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88e3b980-0df4-43e6-bdf1-cab57e2732dc-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.182528 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88e3b980-0df4-43e6-bdf1-cab57e2732dc-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.628829 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkmr9" event={"ID":"88e3b980-0df4-43e6-bdf1-cab57e2732dc","Type":"ContainerDied","Data":"ba1e12932fb07bb4afafdab6a1993ac71c43498f2e76bf016f2096bba7c66303"} Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.628888 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba1e12932fb07bb4afafdab6a1993ac71c43498f2e76bf016f2096bba7c66303" Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.628966 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkmr9" Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.687163 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf"] Oct 07 14:15:54 crc kubenswrapper[4959]: E1007 14:15:54.687494 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88e3b980-0df4-43e6-bdf1-cab57e2732dc" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.687512 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="88e3b980-0df4-43e6-bdf1-cab57e2732dc" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.687670 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="88e3b980-0df4-43e6-bdf1-cab57e2732dc" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.688251 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf" Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.691317 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.691994 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.692948 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-c6vmh" Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.692954 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.711997 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf"] Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.791087 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ng4s9\" (UniqueName: \"kubernetes.io/projected/cbc98733-35d4-477a-825f-3159319f68ad-kube-api-access-ng4s9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf\" (UID: \"cbc98733-35d4-477a-825f-3159319f68ad\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf" Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.791222 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbc98733-35d4-477a-825f-3159319f68ad-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf\" (UID: \"cbc98733-35d4-477a-825f-3159319f68ad\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf" Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.791312 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbc98733-35d4-477a-825f-3159319f68ad-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf\" (UID: \"cbc98733-35d4-477a-825f-3159319f68ad\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf" Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.892268 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbc98733-35d4-477a-825f-3159319f68ad-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf\" (UID: \"cbc98733-35d4-477a-825f-3159319f68ad\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf" Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.892775 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ng4s9\" (UniqueName: \"kubernetes.io/projected/cbc98733-35d4-477a-825f-3159319f68ad-kube-api-access-ng4s9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf\" (UID: \"cbc98733-35d4-477a-825f-3159319f68ad\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf" Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.892942 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbc98733-35d4-477a-825f-3159319f68ad-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf\" (UID: \"cbc98733-35d4-477a-825f-3159319f68ad\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf" Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.896056 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbc98733-35d4-477a-825f-3159319f68ad-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf\" (UID: \"cbc98733-35d4-477a-825f-3159319f68ad\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf" Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.899136 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbc98733-35d4-477a-825f-3159319f68ad-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf\" (UID: \"cbc98733-35d4-477a-825f-3159319f68ad\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf" Oct 07 14:15:54 crc kubenswrapper[4959]: I1007 14:15:54.912474 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ng4s9\" (UniqueName: \"kubernetes.io/projected/cbc98733-35d4-477a-825f-3159319f68ad-kube-api-access-ng4s9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf\" (UID: \"cbc98733-35d4-477a-825f-3159319f68ad\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf" Oct 07 14:15:55 crc kubenswrapper[4959]: I1007 14:15:55.003333 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf" Oct 07 14:15:55 crc kubenswrapper[4959]: I1007 14:15:55.541498 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf"] Oct 07 14:15:55 crc kubenswrapper[4959]: I1007 14:15:55.642288 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf" event={"ID":"cbc98733-35d4-477a-825f-3159319f68ad","Type":"ContainerStarted","Data":"12dc16e40a313830c2d0ebe024ad45c0456f9862a5af9216349d22d15e9c9911"} Oct 07 14:15:57 crc kubenswrapper[4959]: I1007 14:15:57.663603 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf" event={"ID":"cbc98733-35d4-477a-825f-3159319f68ad","Type":"ContainerStarted","Data":"d595881cf5fc11dbf8725f0a0c36208db33f3dd42dc9d45d72caef1faf92d118"} Oct 07 14:15:57 crc kubenswrapper[4959]: I1007 14:15:57.685419 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf" podStartSLOduration=2.261835714 podStartE2EDuration="3.685400937s" podCreationTimestamp="2025-10-07 14:15:54 +0000 UTC" firstStartedPulling="2025-10-07 14:15:55.551729685 +0000 UTC m=+1837.635134010" lastFinishedPulling="2025-10-07 14:15:56.975294898 +0000 UTC m=+1839.058699233" observedRunningTime="2025-10-07 14:15:57.682899184 +0000 UTC m=+1839.766303509" watchObservedRunningTime="2025-10-07 14:15:57.685400937 +0000 UTC m=+1839.768805262" Oct 07 14:16:03 crc kubenswrapper[4959]: I1007 14:16:03.653597 4959 scope.go:117] "RemoveContainer" containerID="8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1" Oct 07 14:16:04 crc kubenswrapper[4959]: I1007 14:16:04.745459 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerStarted","Data":"a084da1c14011d914fafa5905b78b9748fca6b5f32d398b4938e3e75516222b8"} Oct 07 14:16:07 crc kubenswrapper[4959]: I1007 14:16:07.776407 4959 generic.go:334] "Generic (PLEG): container finished" podID="cbc98733-35d4-477a-825f-3159319f68ad" containerID="d595881cf5fc11dbf8725f0a0c36208db33f3dd42dc9d45d72caef1faf92d118" exitCode=0 Oct 07 14:16:07 crc kubenswrapper[4959]: I1007 14:16:07.776507 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf" event={"ID":"cbc98733-35d4-477a-825f-3159319f68ad","Type":"ContainerDied","Data":"d595881cf5fc11dbf8725f0a0c36208db33f3dd42dc9d45d72caef1faf92d118"} Oct 07 14:16:09 crc kubenswrapper[4959]: I1007 14:16:09.214863 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf" Oct 07 14:16:09 crc kubenswrapper[4959]: I1007 14:16:09.297654 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ng4s9\" (UniqueName: \"kubernetes.io/projected/cbc98733-35d4-477a-825f-3159319f68ad-kube-api-access-ng4s9\") pod \"cbc98733-35d4-477a-825f-3159319f68ad\" (UID: \"cbc98733-35d4-477a-825f-3159319f68ad\") " Oct 07 14:16:09 crc kubenswrapper[4959]: I1007 14:16:09.298145 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbc98733-35d4-477a-825f-3159319f68ad-ssh-key\") pod \"cbc98733-35d4-477a-825f-3159319f68ad\" (UID: \"cbc98733-35d4-477a-825f-3159319f68ad\") " Oct 07 14:16:09 crc kubenswrapper[4959]: I1007 14:16:09.298189 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbc98733-35d4-477a-825f-3159319f68ad-inventory\") pod \"cbc98733-35d4-477a-825f-3159319f68ad\" (UID: \"cbc98733-35d4-477a-825f-3159319f68ad\") " Oct 07 14:16:09 crc kubenswrapper[4959]: I1007 14:16:09.303567 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbc98733-35d4-477a-825f-3159319f68ad-kube-api-access-ng4s9" (OuterVolumeSpecName: "kube-api-access-ng4s9") pod "cbc98733-35d4-477a-825f-3159319f68ad" (UID: "cbc98733-35d4-477a-825f-3159319f68ad"). InnerVolumeSpecName "kube-api-access-ng4s9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:16:09 crc kubenswrapper[4959]: I1007 14:16:09.323426 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbc98733-35d4-477a-825f-3159319f68ad-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cbc98733-35d4-477a-825f-3159319f68ad" (UID: "cbc98733-35d4-477a-825f-3159319f68ad"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:16:09 crc kubenswrapper[4959]: I1007 14:16:09.325460 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbc98733-35d4-477a-825f-3159319f68ad-inventory" (OuterVolumeSpecName: "inventory") pod "cbc98733-35d4-477a-825f-3159319f68ad" (UID: "cbc98733-35d4-477a-825f-3159319f68ad"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:16:09 crc kubenswrapper[4959]: I1007 14:16:09.400280 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ng4s9\" (UniqueName: \"kubernetes.io/projected/cbc98733-35d4-477a-825f-3159319f68ad-kube-api-access-ng4s9\") on node \"crc\" DevicePath \"\"" Oct 07 14:16:09 crc kubenswrapper[4959]: I1007 14:16:09.400627 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbc98733-35d4-477a-825f-3159319f68ad-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 14:16:09 crc kubenswrapper[4959]: I1007 14:16:09.400742 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbc98733-35d4-477a-825f-3159319f68ad-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 14:16:09 crc kubenswrapper[4959]: I1007 14:16:09.799441 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf" Oct 07 14:16:09 crc kubenswrapper[4959]: I1007 14:16:09.799470 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf" event={"ID":"cbc98733-35d4-477a-825f-3159319f68ad","Type":"ContainerDied","Data":"12dc16e40a313830c2d0ebe024ad45c0456f9862a5af9216349d22d15e9c9911"} Oct 07 14:16:09 crc kubenswrapper[4959]: I1007 14:16:09.799530 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12dc16e40a313830c2d0ebe024ad45c0456f9862a5af9216349d22d15e9c9911" Oct 07 14:16:14 crc kubenswrapper[4959]: I1007 14:16:14.049322 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-dkgbw"] Oct 07 14:16:14 crc kubenswrapper[4959]: I1007 14:16:14.055912 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-dkgbw"] Oct 07 14:16:14 crc kubenswrapper[4959]: I1007 14:16:14.665984 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adff668e-a674-4ad1-ae8f-a28b2f6bc9e3" path="/var/lib/kubelet/pods/adff668e-a674-4ad1-ae8f-a28b2f6bc9e3/volumes" Oct 07 14:16:29 crc kubenswrapper[4959]: I1007 14:16:29.225509 4959 scope.go:117] "RemoveContainer" containerID="4a4fdfe14783d977fa28e9ac9201e6c4e62084c2ea320ec1660feedafe6fbe6b" Oct 07 14:16:29 crc kubenswrapper[4959]: I1007 14:16:29.287778 4959 scope.go:117] "RemoveContainer" containerID="034f3f30a554074bc27c3bf920535bc8bd69bcef6bf3b7bfc5419fa2bc4ed263" Oct 07 14:16:29 crc kubenswrapper[4959]: I1007 14:16:29.329381 4959 scope.go:117] "RemoveContainer" containerID="d39161cb1f9795c46badb98f2d6b46cebd3cef33dd7e54fd7fe68d89f992726c" Oct 07 14:18:20 crc kubenswrapper[4959]: I1007 14:18:20.630443 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:18:20 crc kubenswrapper[4959]: I1007 14:18:20.630801 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:18:50 crc kubenswrapper[4959]: I1007 14:18:50.630611 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:18:50 crc kubenswrapper[4959]: I1007 14:18:50.631176 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:19:20 crc kubenswrapper[4959]: I1007 14:19:20.630552 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:19:20 crc kubenswrapper[4959]: I1007 14:19:20.631234 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:19:20 crc kubenswrapper[4959]: I1007 14:19:20.631293 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 14:19:20 crc kubenswrapper[4959]: I1007 14:19:20.632086 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a084da1c14011d914fafa5905b78b9748fca6b5f32d398b4938e3e75516222b8"} pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 14:19:20 crc kubenswrapper[4959]: I1007 14:19:20.632225 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" containerID="cri-o://a084da1c14011d914fafa5905b78b9748fca6b5f32d398b4938e3e75516222b8" gracePeriod=600 Oct 07 14:19:21 crc kubenswrapper[4959]: I1007 14:19:21.489145 4959 generic.go:334] "Generic (PLEG): container finished" podID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerID="a084da1c14011d914fafa5905b78b9748fca6b5f32d398b4938e3e75516222b8" exitCode=0 Oct 07 14:19:21 crc kubenswrapper[4959]: I1007 14:19:21.489243 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerDied","Data":"a084da1c14011d914fafa5905b78b9748fca6b5f32d398b4938e3e75516222b8"} Oct 07 14:19:21 crc kubenswrapper[4959]: I1007 14:19:21.489772 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerStarted","Data":"0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d"} Oct 07 14:19:21 crc kubenswrapper[4959]: I1007 14:19:21.489800 4959 scope.go:117] "RemoveContainer" containerID="8dcd76abe0ad86fe47882d9d079e5aa1ae5c31e2edfa12f7ab0746d4f46366e1" Oct 07 14:20:20 crc kubenswrapper[4959]: I1007 14:20:20.909448 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-czl5s"] Oct 07 14:20:20 crc kubenswrapper[4959]: E1007 14:20:20.910752 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbc98733-35d4-477a-825f-3159319f68ad" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 07 14:20:20 crc kubenswrapper[4959]: I1007 14:20:20.910776 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbc98733-35d4-477a-825f-3159319f68ad" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 07 14:20:20 crc kubenswrapper[4959]: I1007 14:20:20.911027 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbc98733-35d4-477a-825f-3159319f68ad" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 07 14:20:20 crc kubenswrapper[4959]: I1007 14:20:20.912731 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-czl5s" Oct 07 14:20:20 crc kubenswrapper[4959]: I1007 14:20:20.928726 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-czl5s"] Oct 07 14:20:21 crc kubenswrapper[4959]: I1007 14:20:21.036458 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfhqf\" (UniqueName: \"kubernetes.io/projected/b5826b27-d56e-4fc9-b2d9-52d1095be937-kube-api-access-pfhqf\") pod \"redhat-marketplace-czl5s\" (UID: \"b5826b27-d56e-4fc9-b2d9-52d1095be937\") " pod="openshift-marketplace/redhat-marketplace-czl5s" Oct 07 14:20:21 crc kubenswrapper[4959]: I1007 14:20:21.036524 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5826b27-d56e-4fc9-b2d9-52d1095be937-utilities\") pod \"redhat-marketplace-czl5s\" (UID: \"b5826b27-d56e-4fc9-b2d9-52d1095be937\") " pod="openshift-marketplace/redhat-marketplace-czl5s" Oct 07 14:20:21 crc kubenswrapper[4959]: I1007 14:20:21.036548 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5826b27-d56e-4fc9-b2d9-52d1095be937-catalog-content\") pod \"redhat-marketplace-czl5s\" (UID: \"b5826b27-d56e-4fc9-b2d9-52d1095be937\") " pod="openshift-marketplace/redhat-marketplace-czl5s" Oct 07 14:20:21 crc kubenswrapper[4959]: I1007 14:20:21.138526 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfhqf\" (UniqueName: \"kubernetes.io/projected/b5826b27-d56e-4fc9-b2d9-52d1095be937-kube-api-access-pfhqf\") pod \"redhat-marketplace-czl5s\" (UID: \"b5826b27-d56e-4fc9-b2d9-52d1095be937\") " pod="openshift-marketplace/redhat-marketplace-czl5s" Oct 07 14:20:21 crc kubenswrapper[4959]: I1007 14:20:21.138593 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5826b27-d56e-4fc9-b2d9-52d1095be937-utilities\") pod \"redhat-marketplace-czl5s\" (UID: \"b5826b27-d56e-4fc9-b2d9-52d1095be937\") " pod="openshift-marketplace/redhat-marketplace-czl5s" Oct 07 14:20:21 crc kubenswrapper[4959]: I1007 14:20:21.138615 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5826b27-d56e-4fc9-b2d9-52d1095be937-catalog-content\") pod \"redhat-marketplace-czl5s\" (UID: \"b5826b27-d56e-4fc9-b2d9-52d1095be937\") " pod="openshift-marketplace/redhat-marketplace-czl5s" Oct 07 14:20:21 crc kubenswrapper[4959]: I1007 14:20:21.139146 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5826b27-d56e-4fc9-b2d9-52d1095be937-catalog-content\") pod \"redhat-marketplace-czl5s\" (UID: \"b5826b27-d56e-4fc9-b2d9-52d1095be937\") " pod="openshift-marketplace/redhat-marketplace-czl5s" Oct 07 14:20:21 crc kubenswrapper[4959]: I1007 14:20:21.139270 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5826b27-d56e-4fc9-b2d9-52d1095be937-utilities\") pod \"redhat-marketplace-czl5s\" (UID: \"b5826b27-d56e-4fc9-b2d9-52d1095be937\") " pod="openshift-marketplace/redhat-marketplace-czl5s" Oct 07 14:20:21 crc kubenswrapper[4959]: I1007 14:20:21.161320 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfhqf\" (UniqueName: \"kubernetes.io/projected/b5826b27-d56e-4fc9-b2d9-52d1095be937-kube-api-access-pfhqf\") pod \"redhat-marketplace-czl5s\" (UID: \"b5826b27-d56e-4fc9-b2d9-52d1095be937\") " pod="openshift-marketplace/redhat-marketplace-czl5s" Oct 07 14:20:21 crc kubenswrapper[4959]: I1007 14:20:21.258017 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-czl5s" Oct 07 14:20:21 crc kubenswrapper[4959]: I1007 14:20:21.735371 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-czl5s"] Oct 07 14:20:22 crc kubenswrapper[4959]: I1007 14:20:22.076541 4959 generic.go:334] "Generic (PLEG): container finished" podID="b5826b27-d56e-4fc9-b2d9-52d1095be937" containerID="9369332a8d0d9ff2bcc44793d6872d915347e1e534ee67ac288677e6d64e6c60" exitCode=0 Oct 07 14:20:22 crc kubenswrapper[4959]: I1007 14:20:22.076695 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-czl5s" event={"ID":"b5826b27-d56e-4fc9-b2d9-52d1095be937","Type":"ContainerDied","Data":"9369332a8d0d9ff2bcc44793d6872d915347e1e534ee67ac288677e6d64e6c60"} Oct 07 14:20:22 crc kubenswrapper[4959]: I1007 14:20:22.077237 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-czl5s" event={"ID":"b5826b27-d56e-4fc9-b2d9-52d1095be937","Type":"ContainerStarted","Data":"6c09c1dbdea3cbc452b0ad9d1355c86486ad9ad67106e6dbdab96714c6713f60"} Oct 07 14:20:22 crc kubenswrapper[4959]: I1007 14:20:22.079518 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 14:20:23 crc kubenswrapper[4959]: I1007 14:20:23.087428 4959 generic.go:334] "Generic (PLEG): container finished" podID="b5826b27-d56e-4fc9-b2d9-52d1095be937" containerID="9a9ddc76908495e5187855a8f58a199005853e78dd7cbd4b5888b358e462c82b" exitCode=0 Oct 07 14:20:23 crc kubenswrapper[4959]: I1007 14:20:23.087492 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-czl5s" event={"ID":"b5826b27-d56e-4fc9-b2d9-52d1095be937","Type":"ContainerDied","Data":"9a9ddc76908495e5187855a8f58a199005853e78dd7cbd4b5888b358e462c82b"} Oct 07 14:20:24 crc kubenswrapper[4959]: I1007 14:20:24.097307 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-czl5s" event={"ID":"b5826b27-d56e-4fc9-b2d9-52d1095be937","Type":"ContainerStarted","Data":"b1cbb9592a84105e7b27972c2d78f76b9f099603f14df0e24aeb94ac833d28ed"} Oct 07 14:20:24 crc kubenswrapper[4959]: I1007 14:20:24.118459 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-czl5s" podStartSLOduration=2.5067768580000003 podStartE2EDuration="4.118439801s" podCreationTimestamp="2025-10-07 14:20:20 +0000 UTC" firstStartedPulling="2025-10-07 14:20:22.079220515 +0000 UTC m=+2104.162624840" lastFinishedPulling="2025-10-07 14:20:23.690883458 +0000 UTC m=+2105.774287783" observedRunningTime="2025-10-07 14:20:24.116489295 +0000 UTC m=+2106.199893620" watchObservedRunningTime="2025-10-07 14:20:24.118439801 +0000 UTC m=+2106.201844126" Oct 07 14:20:29 crc kubenswrapper[4959]: I1007 14:20:29.874759 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xbwtd"] Oct 07 14:20:29 crc kubenswrapper[4959]: I1007 14:20:29.879750 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xbwtd" Oct 07 14:20:29 crc kubenswrapper[4959]: I1007 14:20:29.884394 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xbwtd"] Oct 07 14:20:29 crc kubenswrapper[4959]: I1007 14:20:29.969674 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a1cf027-a138-4b8e-9c60-8325c6103204-utilities\") pod \"certified-operators-xbwtd\" (UID: \"2a1cf027-a138-4b8e-9c60-8325c6103204\") " pod="openshift-marketplace/certified-operators-xbwtd" Oct 07 14:20:29 crc kubenswrapper[4959]: I1007 14:20:29.970033 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5m8q\" (UniqueName: \"kubernetes.io/projected/2a1cf027-a138-4b8e-9c60-8325c6103204-kube-api-access-s5m8q\") pod \"certified-operators-xbwtd\" (UID: \"2a1cf027-a138-4b8e-9c60-8325c6103204\") " pod="openshift-marketplace/certified-operators-xbwtd" Oct 07 14:20:29 crc kubenswrapper[4959]: I1007 14:20:29.970137 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a1cf027-a138-4b8e-9c60-8325c6103204-catalog-content\") pod \"certified-operators-xbwtd\" (UID: \"2a1cf027-a138-4b8e-9c60-8325c6103204\") " pod="openshift-marketplace/certified-operators-xbwtd" Oct 07 14:20:30 crc kubenswrapper[4959]: I1007 14:20:30.073397 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a1cf027-a138-4b8e-9c60-8325c6103204-catalog-content\") pod \"certified-operators-xbwtd\" (UID: \"2a1cf027-a138-4b8e-9c60-8325c6103204\") " pod="openshift-marketplace/certified-operators-xbwtd" Oct 07 14:20:30 crc kubenswrapper[4959]: I1007 14:20:30.073590 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a1cf027-a138-4b8e-9c60-8325c6103204-utilities\") pod \"certified-operators-xbwtd\" (UID: \"2a1cf027-a138-4b8e-9c60-8325c6103204\") " pod="openshift-marketplace/certified-operators-xbwtd" Oct 07 14:20:30 crc kubenswrapper[4959]: I1007 14:20:30.073618 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5m8q\" (UniqueName: \"kubernetes.io/projected/2a1cf027-a138-4b8e-9c60-8325c6103204-kube-api-access-s5m8q\") pod \"certified-operators-xbwtd\" (UID: \"2a1cf027-a138-4b8e-9c60-8325c6103204\") " pod="openshift-marketplace/certified-operators-xbwtd" Oct 07 14:20:30 crc kubenswrapper[4959]: I1007 14:20:30.074320 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a1cf027-a138-4b8e-9c60-8325c6103204-catalog-content\") pod \"certified-operators-xbwtd\" (UID: \"2a1cf027-a138-4b8e-9c60-8325c6103204\") " pod="openshift-marketplace/certified-operators-xbwtd" Oct 07 14:20:30 crc kubenswrapper[4959]: I1007 14:20:30.074366 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a1cf027-a138-4b8e-9c60-8325c6103204-utilities\") pod \"certified-operators-xbwtd\" (UID: \"2a1cf027-a138-4b8e-9c60-8325c6103204\") " pod="openshift-marketplace/certified-operators-xbwtd" Oct 07 14:20:30 crc kubenswrapper[4959]: I1007 14:20:30.102626 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5m8q\" (UniqueName: \"kubernetes.io/projected/2a1cf027-a138-4b8e-9c60-8325c6103204-kube-api-access-s5m8q\") pod \"certified-operators-xbwtd\" (UID: \"2a1cf027-a138-4b8e-9c60-8325c6103204\") " pod="openshift-marketplace/certified-operators-xbwtd" Oct 07 14:20:30 crc kubenswrapper[4959]: I1007 14:20:30.213971 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xbwtd" Oct 07 14:20:30 crc kubenswrapper[4959]: I1007 14:20:30.756727 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xbwtd"] Oct 07 14:20:31 crc kubenswrapper[4959]: I1007 14:20:31.172046 4959 generic.go:334] "Generic (PLEG): container finished" podID="2a1cf027-a138-4b8e-9c60-8325c6103204" containerID="3566750c2a974c351125ede33fc6fc6272d97cfd2c2e78d6c05051d9802d157c" exitCode=0 Oct 07 14:20:31 crc kubenswrapper[4959]: I1007 14:20:31.172332 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xbwtd" event={"ID":"2a1cf027-a138-4b8e-9c60-8325c6103204","Type":"ContainerDied","Data":"3566750c2a974c351125ede33fc6fc6272d97cfd2c2e78d6c05051d9802d157c"} Oct 07 14:20:31 crc kubenswrapper[4959]: I1007 14:20:31.173072 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xbwtd" event={"ID":"2a1cf027-a138-4b8e-9c60-8325c6103204","Type":"ContainerStarted","Data":"e7e7c3635b193609ca01909feeb1a87736125ef8649ba75419cc4d7afa0d38ac"} Oct 07 14:20:31 crc kubenswrapper[4959]: I1007 14:20:31.258651 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-czl5s" Oct 07 14:20:31 crc kubenswrapper[4959]: I1007 14:20:31.259031 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-czl5s" Oct 07 14:20:31 crc kubenswrapper[4959]: I1007 14:20:31.309258 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-czl5s" Oct 07 14:20:32 crc kubenswrapper[4959]: I1007 14:20:32.188578 4959 generic.go:334] "Generic (PLEG): container finished" podID="2a1cf027-a138-4b8e-9c60-8325c6103204" containerID="81b38ce27b2c6a4144b4f746457686889aec5e6efa6f1436c39df52df009b3b7" exitCode=0 Oct 07 14:20:32 crc kubenswrapper[4959]: I1007 14:20:32.188771 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xbwtd" event={"ID":"2a1cf027-a138-4b8e-9c60-8325c6103204","Type":"ContainerDied","Data":"81b38ce27b2c6a4144b4f746457686889aec5e6efa6f1436c39df52df009b3b7"} Oct 07 14:20:32 crc kubenswrapper[4959]: I1007 14:20:32.242377 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-czl5s" Oct 07 14:20:33 crc kubenswrapper[4959]: I1007 14:20:33.200013 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xbwtd" event={"ID":"2a1cf027-a138-4b8e-9c60-8325c6103204","Type":"ContainerStarted","Data":"161effac81806f829cef8c5a4cd4d6b590b1793caf45ef612789a31f095cea9d"} Oct 07 14:20:33 crc kubenswrapper[4959]: I1007 14:20:33.234689 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xbwtd" podStartSLOduration=2.823808062 podStartE2EDuration="4.234672002s" podCreationTimestamp="2025-10-07 14:20:29 +0000 UTC" firstStartedPulling="2025-10-07 14:20:31.174223266 +0000 UTC m=+2113.257627591" lastFinishedPulling="2025-10-07 14:20:32.585087206 +0000 UTC m=+2114.668491531" observedRunningTime="2025-10-07 14:20:33.230911984 +0000 UTC m=+2115.314316309" watchObservedRunningTime="2025-10-07 14:20:33.234672002 +0000 UTC m=+2115.318076327" Oct 07 14:20:33 crc kubenswrapper[4959]: I1007 14:20:33.643234 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-czl5s"] Oct 07 14:20:35 crc kubenswrapper[4959]: I1007 14:20:35.234805 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-czl5s" podUID="b5826b27-d56e-4fc9-b2d9-52d1095be937" containerName="registry-server" containerID="cri-o://b1cbb9592a84105e7b27972c2d78f76b9f099603f14df0e24aeb94ac833d28ed" gracePeriod=2 Oct 07 14:20:35 crc kubenswrapper[4959]: I1007 14:20:35.783192 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-czl5s" Oct 07 14:20:35 crc kubenswrapper[4959]: I1007 14:20:35.801158 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5826b27-d56e-4fc9-b2d9-52d1095be937-utilities\") pod \"b5826b27-d56e-4fc9-b2d9-52d1095be937\" (UID: \"b5826b27-d56e-4fc9-b2d9-52d1095be937\") " Oct 07 14:20:35 crc kubenswrapper[4959]: I1007 14:20:35.801480 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfhqf\" (UniqueName: \"kubernetes.io/projected/b5826b27-d56e-4fc9-b2d9-52d1095be937-kube-api-access-pfhqf\") pod \"b5826b27-d56e-4fc9-b2d9-52d1095be937\" (UID: \"b5826b27-d56e-4fc9-b2d9-52d1095be937\") " Oct 07 14:20:35 crc kubenswrapper[4959]: I1007 14:20:35.801551 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5826b27-d56e-4fc9-b2d9-52d1095be937-catalog-content\") pod \"b5826b27-d56e-4fc9-b2d9-52d1095be937\" (UID: \"b5826b27-d56e-4fc9-b2d9-52d1095be937\") " Oct 07 14:20:35 crc kubenswrapper[4959]: I1007 14:20:35.804182 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5826b27-d56e-4fc9-b2d9-52d1095be937-utilities" (OuterVolumeSpecName: "utilities") pod "b5826b27-d56e-4fc9-b2d9-52d1095be937" (UID: "b5826b27-d56e-4fc9-b2d9-52d1095be937"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:20:35 crc kubenswrapper[4959]: I1007 14:20:35.810138 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5826b27-d56e-4fc9-b2d9-52d1095be937-kube-api-access-pfhqf" (OuterVolumeSpecName: "kube-api-access-pfhqf") pod "b5826b27-d56e-4fc9-b2d9-52d1095be937" (UID: "b5826b27-d56e-4fc9-b2d9-52d1095be937"). InnerVolumeSpecName "kube-api-access-pfhqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:20:35 crc kubenswrapper[4959]: I1007 14:20:35.825535 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5826b27-d56e-4fc9-b2d9-52d1095be937-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b5826b27-d56e-4fc9-b2d9-52d1095be937" (UID: "b5826b27-d56e-4fc9-b2d9-52d1095be937"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:20:35 crc kubenswrapper[4959]: I1007 14:20:35.903989 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5826b27-d56e-4fc9-b2d9-52d1095be937-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 14:20:35 crc kubenswrapper[4959]: I1007 14:20:35.904030 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5826b27-d56e-4fc9-b2d9-52d1095be937-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 14:20:35 crc kubenswrapper[4959]: I1007 14:20:35.904040 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfhqf\" (UniqueName: \"kubernetes.io/projected/b5826b27-d56e-4fc9-b2d9-52d1095be937-kube-api-access-pfhqf\") on node \"crc\" DevicePath \"\"" Oct 07 14:20:36 crc kubenswrapper[4959]: I1007 14:20:36.249885 4959 generic.go:334] "Generic (PLEG): container finished" podID="b5826b27-d56e-4fc9-b2d9-52d1095be937" containerID="b1cbb9592a84105e7b27972c2d78f76b9f099603f14df0e24aeb94ac833d28ed" exitCode=0 Oct 07 14:20:36 crc kubenswrapper[4959]: I1007 14:20:36.249958 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-czl5s" event={"ID":"b5826b27-d56e-4fc9-b2d9-52d1095be937","Type":"ContainerDied","Data":"b1cbb9592a84105e7b27972c2d78f76b9f099603f14df0e24aeb94ac833d28ed"} Oct 07 14:20:36 crc kubenswrapper[4959]: I1007 14:20:36.250029 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-czl5s" Oct 07 14:20:36 crc kubenswrapper[4959]: I1007 14:20:36.250074 4959 scope.go:117] "RemoveContainer" containerID="b1cbb9592a84105e7b27972c2d78f76b9f099603f14df0e24aeb94ac833d28ed" Oct 07 14:20:36 crc kubenswrapper[4959]: I1007 14:20:36.250055 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-czl5s" event={"ID":"b5826b27-d56e-4fc9-b2d9-52d1095be937","Type":"ContainerDied","Data":"6c09c1dbdea3cbc452b0ad9d1355c86486ad9ad67106e6dbdab96714c6713f60"} Oct 07 14:20:36 crc kubenswrapper[4959]: I1007 14:20:36.283127 4959 scope.go:117] "RemoveContainer" containerID="9a9ddc76908495e5187855a8f58a199005853e78dd7cbd4b5888b358e462c82b" Oct 07 14:20:36 crc kubenswrapper[4959]: I1007 14:20:36.300971 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-czl5s"] Oct 07 14:20:36 crc kubenswrapper[4959]: I1007 14:20:36.311430 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-czl5s"] Oct 07 14:20:36 crc kubenswrapper[4959]: I1007 14:20:36.316517 4959 scope.go:117] "RemoveContainer" containerID="9369332a8d0d9ff2bcc44793d6872d915347e1e534ee67ac288677e6d64e6c60" Oct 07 14:20:36 crc kubenswrapper[4959]: I1007 14:20:36.356880 4959 scope.go:117] "RemoveContainer" containerID="b1cbb9592a84105e7b27972c2d78f76b9f099603f14df0e24aeb94ac833d28ed" Oct 07 14:20:36 crc kubenswrapper[4959]: E1007 14:20:36.357479 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1cbb9592a84105e7b27972c2d78f76b9f099603f14df0e24aeb94ac833d28ed\": container with ID starting with b1cbb9592a84105e7b27972c2d78f76b9f099603f14df0e24aeb94ac833d28ed not found: ID does not exist" containerID="b1cbb9592a84105e7b27972c2d78f76b9f099603f14df0e24aeb94ac833d28ed" Oct 07 14:20:36 crc kubenswrapper[4959]: I1007 14:20:36.357537 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1cbb9592a84105e7b27972c2d78f76b9f099603f14df0e24aeb94ac833d28ed"} err="failed to get container status \"b1cbb9592a84105e7b27972c2d78f76b9f099603f14df0e24aeb94ac833d28ed\": rpc error: code = NotFound desc = could not find container \"b1cbb9592a84105e7b27972c2d78f76b9f099603f14df0e24aeb94ac833d28ed\": container with ID starting with b1cbb9592a84105e7b27972c2d78f76b9f099603f14df0e24aeb94ac833d28ed not found: ID does not exist" Oct 07 14:20:36 crc kubenswrapper[4959]: I1007 14:20:36.357577 4959 scope.go:117] "RemoveContainer" containerID="9a9ddc76908495e5187855a8f58a199005853e78dd7cbd4b5888b358e462c82b" Oct 07 14:20:36 crc kubenswrapper[4959]: E1007 14:20:36.358191 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a9ddc76908495e5187855a8f58a199005853e78dd7cbd4b5888b358e462c82b\": container with ID starting with 9a9ddc76908495e5187855a8f58a199005853e78dd7cbd4b5888b358e462c82b not found: ID does not exist" containerID="9a9ddc76908495e5187855a8f58a199005853e78dd7cbd4b5888b358e462c82b" Oct 07 14:20:36 crc kubenswrapper[4959]: I1007 14:20:36.358243 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a9ddc76908495e5187855a8f58a199005853e78dd7cbd4b5888b358e462c82b"} err="failed to get container status \"9a9ddc76908495e5187855a8f58a199005853e78dd7cbd4b5888b358e462c82b\": rpc error: code = NotFound desc = could not find container \"9a9ddc76908495e5187855a8f58a199005853e78dd7cbd4b5888b358e462c82b\": container with ID starting with 9a9ddc76908495e5187855a8f58a199005853e78dd7cbd4b5888b358e462c82b not found: ID does not exist" Oct 07 14:20:36 crc kubenswrapper[4959]: I1007 14:20:36.358280 4959 scope.go:117] "RemoveContainer" containerID="9369332a8d0d9ff2bcc44793d6872d915347e1e534ee67ac288677e6d64e6c60" Oct 07 14:20:36 crc kubenswrapper[4959]: E1007 14:20:36.358645 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9369332a8d0d9ff2bcc44793d6872d915347e1e534ee67ac288677e6d64e6c60\": container with ID starting with 9369332a8d0d9ff2bcc44793d6872d915347e1e534ee67ac288677e6d64e6c60 not found: ID does not exist" containerID="9369332a8d0d9ff2bcc44793d6872d915347e1e534ee67ac288677e6d64e6c60" Oct 07 14:20:36 crc kubenswrapper[4959]: I1007 14:20:36.358717 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9369332a8d0d9ff2bcc44793d6872d915347e1e534ee67ac288677e6d64e6c60"} err="failed to get container status \"9369332a8d0d9ff2bcc44793d6872d915347e1e534ee67ac288677e6d64e6c60\": rpc error: code = NotFound desc = could not find container \"9369332a8d0d9ff2bcc44793d6872d915347e1e534ee67ac288677e6d64e6c60\": container with ID starting with 9369332a8d0d9ff2bcc44793d6872d915347e1e534ee67ac288677e6d64e6c60 not found: ID does not exist" Oct 07 14:20:36 crc kubenswrapper[4959]: I1007 14:20:36.667876 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5826b27-d56e-4fc9-b2d9-52d1095be937" path="/var/lib/kubelet/pods/b5826b27-d56e-4fc9-b2d9-52d1095be937/volumes" Oct 07 14:20:40 crc kubenswrapper[4959]: I1007 14:20:40.214901 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xbwtd" Oct 07 14:20:40 crc kubenswrapper[4959]: I1007 14:20:40.215981 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xbwtd" Oct 07 14:20:40 crc kubenswrapper[4959]: I1007 14:20:40.275175 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xbwtd" Oct 07 14:20:40 crc kubenswrapper[4959]: I1007 14:20:40.351376 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xbwtd" Oct 07 14:20:40 crc kubenswrapper[4959]: I1007 14:20:40.644331 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xbwtd"] Oct 07 14:20:42 crc kubenswrapper[4959]: I1007 14:20:42.317531 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xbwtd" podUID="2a1cf027-a138-4b8e-9c60-8325c6103204" containerName="registry-server" containerID="cri-o://161effac81806f829cef8c5a4cd4d6b590b1793caf45ef612789a31f095cea9d" gracePeriod=2 Oct 07 14:20:42 crc kubenswrapper[4959]: I1007 14:20:42.844537 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xbwtd" Oct 07 14:20:42 crc kubenswrapper[4959]: I1007 14:20:42.969637 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a1cf027-a138-4b8e-9c60-8325c6103204-utilities\") pod \"2a1cf027-a138-4b8e-9c60-8325c6103204\" (UID: \"2a1cf027-a138-4b8e-9c60-8325c6103204\") " Oct 07 14:20:42 crc kubenswrapper[4959]: I1007 14:20:42.969774 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5m8q\" (UniqueName: \"kubernetes.io/projected/2a1cf027-a138-4b8e-9c60-8325c6103204-kube-api-access-s5m8q\") pod \"2a1cf027-a138-4b8e-9c60-8325c6103204\" (UID: \"2a1cf027-a138-4b8e-9c60-8325c6103204\") " Oct 07 14:20:42 crc kubenswrapper[4959]: I1007 14:20:42.969934 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a1cf027-a138-4b8e-9c60-8325c6103204-catalog-content\") pod \"2a1cf027-a138-4b8e-9c60-8325c6103204\" (UID: \"2a1cf027-a138-4b8e-9c60-8325c6103204\") " Oct 07 14:20:42 crc kubenswrapper[4959]: I1007 14:20:42.971276 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a1cf027-a138-4b8e-9c60-8325c6103204-utilities" (OuterVolumeSpecName: "utilities") pod "2a1cf027-a138-4b8e-9c60-8325c6103204" (UID: "2a1cf027-a138-4b8e-9c60-8325c6103204"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:20:42 crc kubenswrapper[4959]: I1007 14:20:42.971609 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a1cf027-a138-4b8e-9c60-8325c6103204-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 14:20:42 crc kubenswrapper[4959]: I1007 14:20:42.981512 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a1cf027-a138-4b8e-9c60-8325c6103204-kube-api-access-s5m8q" (OuterVolumeSpecName: "kube-api-access-s5m8q") pod "2a1cf027-a138-4b8e-9c60-8325c6103204" (UID: "2a1cf027-a138-4b8e-9c60-8325c6103204"). InnerVolumeSpecName "kube-api-access-s5m8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:20:43 crc kubenswrapper[4959]: I1007 14:20:43.073387 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5m8q\" (UniqueName: \"kubernetes.io/projected/2a1cf027-a138-4b8e-9c60-8325c6103204-kube-api-access-s5m8q\") on node \"crc\" DevicePath \"\"" Oct 07 14:20:43 crc kubenswrapper[4959]: I1007 14:20:43.327577 4959 generic.go:334] "Generic (PLEG): container finished" podID="2a1cf027-a138-4b8e-9c60-8325c6103204" containerID="161effac81806f829cef8c5a4cd4d6b590b1793caf45ef612789a31f095cea9d" exitCode=0 Oct 07 14:20:43 crc kubenswrapper[4959]: I1007 14:20:43.327629 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xbwtd" event={"ID":"2a1cf027-a138-4b8e-9c60-8325c6103204","Type":"ContainerDied","Data":"161effac81806f829cef8c5a4cd4d6b590b1793caf45ef612789a31f095cea9d"} Oct 07 14:20:43 crc kubenswrapper[4959]: I1007 14:20:43.327668 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xbwtd" Oct 07 14:20:43 crc kubenswrapper[4959]: I1007 14:20:43.327685 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xbwtd" event={"ID":"2a1cf027-a138-4b8e-9c60-8325c6103204","Type":"ContainerDied","Data":"e7e7c3635b193609ca01909feeb1a87736125ef8649ba75419cc4d7afa0d38ac"} Oct 07 14:20:43 crc kubenswrapper[4959]: I1007 14:20:43.327713 4959 scope.go:117] "RemoveContainer" containerID="161effac81806f829cef8c5a4cd4d6b590b1793caf45ef612789a31f095cea9d" Oct 07 14:20:43 crc kubenswrapper[4959]: I1007 14:20:43.362149 4959 scope.go:117] "RemoveContainer" containerID="81b38ce27b2c6a4144b4f746457686889aec5e6efa6f1436c39df52df009b3b7" Oct 07 14:20:43 crc kubenswrapper[4959]: I1007 14:20:43.392695 4959 scope.go:117] "RemoveContainer" containerID="3566750c2a974c351125ede33fc6fc6272d97cfd2c2e78d6c05051d9802d157c" Oct 07 14:20:43 crc kubenswrapper[4959]: I1007 14:20:43.449712 4959 scope.go:117] "RemoveContainer" containerID="161effac81806f829cef8c5a4cd4d6b590b1793caf45ef612789a31f095cea9d" Oct 07 14:20:43 crc kubenswrapper[4959]: E1007 14:20:43.451270 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"161effac81806f829cef8c5a4cd4d6b590b1793caf45ef612789a31f095cea9d\": container with ID starting with 161effac81806f829cef8c5a4cd4d6b590b1793caf45ef612789a31f095cea9d not found: ID does not exist" containerID="161effac81806f829cef8c5a4cd4d6b590b1793caf45ef612789a31f095cea9d" Oct 07 14:20:43 crc kubenswrapper[4959]: I1007 14:20:43.451331 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"161effac81806f829cef8c5a4cd4d6b590b1793caf45ef612789a31f095cea9d"} err="failed to get container status \"161effac81806f829cef8c5a4cd4d6b590b1793caf45ef612789a31f095cea9d\": rpc error: code = NotFound desc = could not find container \"161effac81806f829cef8c5a4cd4d6b590b1793caf45ef612789a31f095cea9d\": container with ID starting with 161effac81806f829cef8c5a4cd4d6b590b1793caf45ef612789a31f095cea9d not found: ID does not exist" Oct 07 14:20:43 crc kubenswrapper[4959]: I1007 14:20:43.451362 4959 scope.go:117] "RemoveContainer" containerID="81b38ce27b2c6a4144b4f746457686889aec5e6efa6f1436c39df52df009b3b7" Oct 07 14:20:43 crc kubenswrapper[4959]: E1007 14:20:43.451979 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81b38ce27b2c6a4144b4f746457686889aec5e6efa6f1436c39df52df009b3b7\": container with ID starting with 81b38ce27b2c6a4144b4f746457686889aec5e6efa6f1436c39df52df009b3b7 not found: ID does not exist" containerID="81b38ce27b2c6a4144b4f746457686889aec5e6efa6f1436c39df52df009b3b7" Oct 07 14:20:43 crc kubenswrapper[4959]: I1007 14:20:43.452031 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81b38ce27b2c6a4144b4f746457686889aec5e6efa6f1436c39df52df009b3b7"} err="failed to get container status \"81b38ce27b2c6a4144b4f746457686889aec5e6efa6f1436c39df52df009b3b7\": rpc error: code = NotFound desc = could not find container \"81b38ce27b2c6a4144b4f746457686889aec5e6efa6f1436c39df52df009b3b7\": container with ID starting with 81b38ce27b2c6a4144b4f746457686889aec5e6efa6f1436c39df52df009b3b7 not found: ID does not exist" Oct 07 14:20:43 crc kubenswrapper[4959]: I1007 14:20:43.452047 4959 scope.go:117] "RemoveContainer" containerID="3566750c2a974c351125ede33fc6fc6272d97cfd2c2e78d6c05051d9802d157c" Oct 07 14:20:43 crc kubenswrapper[4959]: E1007 14:20:43.452705 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3566750c2a974c351125ede33fc6fc6272d97cfd2c2e78d6c05051d9802d157c\": container with ID starting with 3566750c2a974c351125ede33fc6fc6272d97cfd2c2e78d6c05051d9802d157c not found: ID does not exist" containerID="3566750c2a974c351125ede33fc6fc6272d97cfd2c2e78d6c05051d9802d157c" Oct 07 14:20:43 crc kubenswrapper[4959]: I1007 14:20:43.452734 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3566750c2a974c351125ede33fc6fc6272d97cfd2c2e78d6c05051d9802d157c"} err="failed to get container status \"3566750c2a974c351125ede33fc6fc6272d97cfd2c2e78d6c05051d9802d157c\": rpc error: code = NotFound desc = could not find container \"3566750c2a974c351125ede33fc6fc6272d97cfd2c2e78d6c05051d9802d157c\": container with ID starting with 3566750c2a974c351125ede33fc6fc6272d97cfd2c2e78d6c05051d9802d157c not found: ID does not exist" Oct 07 14:20:43 crc kubenswrapper[4959]: I1007 14:20:43.714581 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a1cf027-a138-4b8e-9c60-8325c6103204-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2a1cf027-a138-4b8e-9c60-8325c6103204" (UID: "2a1cf027-a138-4b8e-9c60-8325c6103204"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:20:43 crc kubenswrapper[4959]: I1007 14:20:43.787928 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a1cf027-a138-4b8e-9c60-8325c6103204-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 14:20:43 crc kubenswrapper[4959]: I1007 14:20:43.990291 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xbwtd"] Oct 07 14:20:44 crc kubenswrapper[4959]: I1007 14:20:44.005841 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xbwtd"] Oct 07 14:20:44 crc kubenswrapper[4959]: I1007 14:20:44.668151 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a1cf027-a138-4b8e-9c60-8325c6103204" path="/var/lib/kubelet/pods/2a1cf027-a138-4b8e-9c60-8325c6103204/volumes" Oct 07 14:20:49 crc kubenswrapper[4959]: I1007 14:20:49.078417 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nwksc"] Oct 07 14:20:49 crc kubenswrapper[4959]: E1007 14:20:49.079187 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5826b27-d56e-4fc9-b2d9-52d1095be937" containerName="registry-server" Oct 07 14:20:49 crc kubenswrapper[4959]: I1007 14:20:49.079206 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5826b27-d56e-4fc9-b2d9-52d1095be937" containerName="registry-server" Oct 07 14:20:49 crc kubenswrapper[4959]: E1007 14:20:49.079240 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a1cf027-a138-4b8e-9c60-8325c6103204" containerName="extract-utilities" Oct 07 14:20:49 crc kubenswrapper[4959]: I1007 14:20:49.079249 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a1cf027-a138-4b8e-9c60-8325c6103204" containerName="extract-utilities" Oct 07 14:20:49 crc kubenswrapper[4959]: E1007 14:20:49.079269 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a1cf027-a138-4b8e-9c60-8325c6103204" containerName="registry-server" Oct 07 14:20:49 crc kubenswrapper[4959]: I1007 14:20:49.079279 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a1cf027-a138-4b8e-9c60-8325c6103204" containerName="registry-server" Oct 07 14:20:49 crc kubenswrapper[4959]: E1007 14:20:49.079301 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5826b27-d56e-4fc9-b2d9-52d1095be937" containerName="extract-utilities" Oct 07 14:20:49 crc kubenswrapper[4959]: I1007 14:20:49.079311 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5826b27-d56e-4fc9-b2d9-52d1095be937" containerName="extract-utilities" Oct 07 14:20:49 crc kubenswrapper[4959]: E1007 14:20:49.079325 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a1cf027-a138-4b8e-9c60-8325c6103204" containerName="extract-content" Oct 07 14:20:49 crc kubenswrapper[4959]: I1007 14:20:49.079332 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a1cf027-a138-4b8e-9c60-8325c6103204" containerName="extract-content" Oct 07 14:20:49 crc kubenswrapper[4959]: E1007 14:20:49.079347 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5826b27-d56e-4fc9-b2d9-52d1095be937" containerName="extract-content" Oct 07 14:20:49 crc kubenswrapper[4959]: I1007 14:20:49.079354 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5826b27-d56e-4fc9-b2d9-52d1095be937" containerName="extract-content" Oct 07 14:20:49 crc kubenswrapper[4959]: I1007 14:20:49.079571 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5826b27-d56e-4fc9-b2d9-52d1095be937" containerName="registry-server" Oct 07 14:20:49 crc kubenswrapper[4959]: I1007 14:20:49.079599 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a1cf027-a138-4b8e-9c60-8325c6103204" containerName="registry-server" Oct 07 14:20:49 crc kubenswrapper[4959]: I1007 14:20:49.081353 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nwksc" Oct 07 14:20:49 crc kubenswrapper[4959]: I1007 14:20:49.087975 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nwksc"] Oct 07 14:20:49 crc kubenswrapper[4959]: I1007 14:20:49.090313 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzbm7\" (UniqueName: \"kubernetes.io/projected/fad6288b-9766-448d-9f91-374bc79257dd-kube-api-access-jzbm7\") pod \"redhat-operators-nwksc\" (UID: \"fad6288b-9766-448d-9f91-374bc79257dd\") " pod="openshift-marketplace/redhat-operators-nwksc" Oct 07 14:20:49 crc kubenswrapper[4959]: I1007 14:20:49.090430 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fad6288b-9766-448d-9f91-374bc79257dd-utilities\") pod \"redhat-operators-nwksc\" (UID: \"fad6288b-9766-448d-9f91-374bc79257dd\") " pod="openshift-marketplace/redhat-operators-nwksc" Oct 07 14:20:49 crc kubenswrapper[4959]: I1007 14:20:49.090467 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fad6288b-9766-448d-9f91-374bc79257dd-catalog-content\") pod \"redhat-operators-nwksc\" (UID: \"fad6288b-9766-448d-9f91-374bc79257dd\") " pod="openshift-marketplace/redhat-operators-nwksc" Oct 07 14:20:49 crc kubenswrapper[4959]: I1007 14:20:49.191982 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzbm7\" (UniqueName: \"kubernetes.io/projected/fad6288b-9766-448d-9f91-374bc79257dd-kube-api-access-jzbm7\") pod \"redhat-operators-nwksc\" (UID: \"fad6288b-9766-448d-9f91-374bc79257dd\") " pod="openshift-marketplace/redhat-operators-nwksc" Oct 07 14:20:49 crc kubenswrapper[4959]: I1007 14:20:49.192112 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fad6288b-9766-448d-9f91-374bc79257dd-utilities\") pod \"redhat-operators-nwksc\" (UID: \"fad6288b-9766-448d-9f91-374bc79257dd\") " pod="openshift-marketplace/redhat-operators-nwksc" Oct 07 14:20:49 crc kubenswrapper[4959]: I1007 14:20:49.192147 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fad6288b-9766-448d-9f91-374bc79257dd-catalog-content\") pod \"redhat-operators-nwksc\" (UID: \"fad6288b-9766-448d-9f91-374bc79257dd\") " pod="openshift-marketplace/redhat-operators-nwksc" Oct 07 14:20:49 crc kubenswrapper[4959]: I1007 14:20:49.192820 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fad6288b-9766-448d-9f91-374bc79257dd-utilities\") pod \"redhat-operators-nwksc\" (UID: \"fad6288b-9766-448d-9f91-374bc79257dd\") " pod="openshift-marketplace/redhat-operators-nwksc" Oct 07 14:20:49 crc kubenswrapper[4959]: I1007 14:20:49.192913 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fad6288b-9766-448d-9f91-374bc79257dd-catalog-content\") pod \"redhat-operators-nwksc\" (UID: \"fad6288b-9766-448d-9f91-374bc79257dd\") " pod="openshift-marketplace/redhat-operators-nwksc" Oct 07 14:20:49 crc kubenswrapper[4959]: I1007 14:20:49.229837 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzbm7\" (UniqueName: \"kubernetes.io/projected/fad6288b-9766-448d-9f91-374bc79257dd-kube-api-access-jzbm7\") pod \"redhat-operators-nwksc\" (UID: \"fad6288b-9766-448d-9f91-374bc79257dd\") " pod="openshift-marketplace/redhat-operators-nwksc" Oct 07 14:20:49 crc kubenswrapper[4959]: I1007 14:20:49.407449 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nwksc" Oct 07 14:20:49 crc kubenswrapper[4959]: I1007 14:20:49.873331 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nwksc"] Oct 07 14:20:50 crc kubenswrapper[4959]: I1007 14:20:50.397120 4959 generic.go:334] "Generic (PLEG): container finished" podID="fad6288b-9766-448d-9f91-374bc79257dd" containerID="1360d2e1c0977ca6158531cc5726f0bfb4a0be4c46d55bf75db0e96c89ce2f4b" exitCode=0 Oct 07 14:20:50 crc kubenswrapper[4959]: I1007 14:20:50.397183 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nwksc" event={"ID":"fad6288b-9766-448d-9f91-374bc79257dd","Type":"ContainerDied","Data":"1360d2e1c0977ca6158531cc5726f0bfb4a0be4c46d55bf75db0e96c89ce2f4b"} Oct 07 14:20:50 crc kubenswrapper[4959]: I1007 14:20:50.397381 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nwksc" event={"ID":"fad6288b-9766-448d-9f91-374bc79257dd","Type":"ContainerStarted","Data":"8b1d9d995a76d4186c146f31a9a45f5b674d9a6206c89a572c9ca26700969d5f"} Oct 07 14:20:57 crc kubenswrapper[4959]: I1007 14:20:57.483162 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nwksc" event={"ID":"fad6288b-9766-448d-9f91-374bc79257dd","Type":"ContainerStarted","Data":"7c28e4501f6143d9a431c9062d327051d5d1eb10a33bab38661542680c104d28"} Oct 07 14:20:59 crc kubenswrapper[4959]: I1007 14:20:59.508045 4959 generic.go:334] "Generic (PLEG): container finished" podID="fad6288b-9766-448d-9f91-374bc79257dd" containerID="7c28e4501f6143d9a431c9062d327051d5d1eb10a33bab38661542680c104d28" exitCode=0 Oct 07 14:20:59 crc kubenswrapper[4959]: I1007 14:20:59.508126 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nwksc" event={"ID":"fad6288b-9766-448d-9f91-374bc79257dd","Type":"ContainerDied","Data":"7c28e4501f6143d9a431c9062d327051d5d1eb10a33bab38661542680c104d28"} Oct 07 14:21:00 crc kubenswrapper[4959]: I1007 14:21:00.518702 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nwksc" event={"ID":"fad6288b-9766-448d-9f91-374bc79257dd","Type":"ContainerStarted","Data":"6424b528e2636d9fd5db8a284452d6d3698a0710c9d3984ee3447099dbf624a2"} Oct 07 14:21:00 crc kubenswrapper[4959]: I1007 14:21:00.543460 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nwksc" podStartSLOduration=2.045901115 podStartE2EDuration="11.543439107s" podCreationTimestamp="2025-10-07 14:20:49 +0000 UTC" firstStartedPulling="2025-10-07 14:20:50.399138279 +0000 UTC m=+2132.482542594" lastFinishedPulling="2025-10-07 14:20:59.896676261 +0000 UTC m=+2141.980080586" observedRunningTime="2025-10-07 14:21:00.534626583 +0000 UTC m=+2142.618030908" watchObservedRunningTime="2025-10-07 14:21:00.543439107 +0000 UTC m=+2142.626843432" Oct 07 14:21:09 crc kubenswrapper[4959]: I1007 14:21:09.407761 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nwksc" Oct 07 14:21:09 crc kubenswrapper[4959]: I1007 14:21:09.408345 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nwksc" Oct 07 14:21:09 crc kubenswrapper[4959]: I1007 14:21:09.464347 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nwksc" Oct 07 14:21:09 crc kubenswrapper[4959]: I1007 14:21:09.656140 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nwksc" Oct 07 14:21:09 crc kubenswrapper[4959]: I1007 14:21:09.738400 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nwksc"] Oct 07 14:21:09 crc kubenswrapper[4959]: I1007 14:21:09.793849 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jhqgr"] Oct 07 14:21:09 crc kubenswrapper[4959]: I1007 14:21:09.794192 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jhqgr" podUID="2f98e86e-cc03-4344-82d6-897514556718" containerName="registry-server" containerID="cri-o://a24c3a555f2faa4755cf004da20929226b1ea34cc3fbaac4cd616ea3be782569" gracePeriod=2 Oct 07 14:21:10 crc kubenswrapper[4959]: I1007 14:21:10.619886 4959 generic.go:334] "Generic (PLEG): container finished" podID="2f98e86e-cc03-4344-82d6-897514556718" containerID="a24c3a555f2faa4755cf004da20929226b1ea34cc3fbaac4cd616ea3be782569" exitCode=0 Oct 07 14:21:10 crc kubenswrapper[4959]: I1007 14:21:10.621157 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhqgr" event={"ID":"2f98e86e-cc03-4344-82d6-897514556718","Type":"ContainerDied","Data":"a24c3a555f2faa4755cf004da20929226b1ea34cc3fbaac4cd616ea3be782569"} Oct 07 14:21:10 crc kubenswrapper[4959]: I1007 14:21:10.983717 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jhqgr" Oct 07 14:21:11 crc kubenswrapper[4959]: I1007 14:21:11.046525 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5g9n\" (UniqueName: \"kubernetes.io/projected/2f98e86e-cc03-4344-82d6-897514556718-kube-api-access-w5g9n\") pod \"2f98e86e-cc03-4344-82d6-897514556718\" (UID: \"2f98e86e-cc03-4344-82d6-897514556718\") " Oct 07 14:21:11 crc kubenswrapper[4959]: I1007 14:21:11.046739 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f98e86e-cc03-4344-82d6-897514556718-catalog-content\") pod \"2f98e86e-cc03-4344-82d6-897514556718\" (UID: \"2f98e86e-cc03-4344-82d6-897514556718\") " Oct 07 14:21:11 crc kubenswrapper[4959]: I1007 14:21:11.046799 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f98e86e-cc03-4344-82d6-897514556718-utilities\") pod \"2f98e86e-cc03-4344-82d6-897514556718\" (UID: \"2f98e86e-cc03-4344-82d6-897514556718\") " Oct 07 14:21:11 crc kubenswrapper[4959]: I1007 14:21:11.047743 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f98e86e-cc03-4344-82d6-897514556718-utilities" (OuterVolumeSpecName: "utilities") pod "2f98e86e-cc03-4344-82d6-897514556718" (UID: "2f98e86e-cc03-4344-82d6-897514556718"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:21:11 crc kubenswrapper[4959]: I1007 14:21:11.056839 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f98e86e-cc03-4344-82d6-897514556718-kube-api-access-w5g9n" (OuterVolumeSpecName: "kube-api-access-w5g9n") pod "2f98e86e-cc03-4344-82d6-897514556718" (UID: "2f98e86e-cc03-4344-82d6-897514556718"). InnerVolumeSpecName "kube-api-access-w5g9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:21:11 crc kubenswrapper[4959]: I1007 14:21:11.124942 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f98e86e-cc03-4344-82d6-897514556718-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2f98e86e-cc03-4344-82d6-897514556718" (UID: "2f98e86e-cc03-4344-82d6-897514556718"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:21:11 crc kubenswrapper[4959]: I1007 14:21:11.149680 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5g9n\" (UniqueName: \"kubernetes.io/projected/2f98e86e-cc03-4344-82d6-897514556718-kube-api-access-w5g9n\") on node \"crc\" DevicePath \"\"" Oct 07 14:21:11 crc kubenswrapper[4959]: I1007 14:21:11.150123 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f98e86e-cc03-4344-82d6-897514556718-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 14:21:11 crc kubenswrapper[4959]: I1007 14:21:11.150202 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f98e86e-cc03-4344-82d6-897514556718-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 14:21:11 crc kubenswrapper[4959]: I1007 14:21:11.631064 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhqgr" event={"ID":"2f98e86e-cc03-4344-82d6-897514556718","Type":"ContainerDied","Data":"42ad41d343ea16415faaf4b64918d0d04cb430cc56aa29853e571b4c00ac768d"} Oct 07 14:21:11 crc kubenswrapper[4959]: I1007 14:21:11.631134 4959 scope.go:117] "RemoveContainer" containerID="a24c3a555f2faa4755cf004da20929226b1ea34cc3fbaac4cd616ea3be782569" Oct 07 14:21:11 crc kubenswrapper[4959]: I1007 14:21:11.631084 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jhqgr" Oct 07 14:21:11 crc kubenswrapper[4959]: I1007 14:21:11.658147 4959 scope.go:117] "RemoveContainer" containerID="79d577eb44ebdadd81340ebed939ea537a89e0860a636dfc426a659f16b3abad" Oct 07 14:21:11 crc kubenswrapper[4959]: I1007 14:21:11.680037 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jhqgr"] Oct 07 14:21:11 crc kubenswrapper[4959]: I1007 14:21:11.688117 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jhqgr"] Oct 07 14:21:11 crc kubenswrapper[4959]: I1007 14:21:11.710623 4959 scope.go:117] "RemoveContainer" containerID="94d62cb480fba8ae245fb9eac6e93f7e178a06eb156c055d7ba5a6a1ef3bd06a" Oct 07 14:21:11 crc kubenswrapper[4959]: E1007 14:21:11.769534 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f98e86e_cc03_4344_82d6_897514556718.slice\": RecentStats: unable to find data in memory cache]" Oct 07 14:21:12 crc kubenswrapper[4959]: I1007 14:21:12.664216 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f98e86e-cc03-4344-82d6-897514556718" path="/var/lib/kubelet/pods/2f98e86e-cc03-4344-82d6-897514556718/volumes" Oct 07 14:21:20 crc kubenswrapper[4959]: I1007 14:21:20.630532 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:21:20 crc kubenswrapper[4959]: I1007 14:21:20.631113 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:21:50 crc kubenswrapper[4959]: I1007 14:21:50.630159 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:21:50 crc kubenswrapper[4959]: I1007 14:21:50.630633 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:22:12 crc kubenswrapper[4959]: E1007 14:22:12.700243 4959 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.222:53586->38.102.83.222:33441: write tcp 38.102.83.222:53586->38.102.83.222:33441: write: broken pipe Oct 07 14:22:19 crc kubenswrapper[4959]: E1007 14:22:19.814332 4959 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.222:51422->38.102.83.222:33441: write tcp 38.102.83.222:51422->38.102.83.222:33441: write: broken pipe Oct 07 14:22:20 crc kubenswrapper[4959]: I1007 14:22:20.630065 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:22:20 crc kubenswrapper[4959]: I1007 14:22:20.630635 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:22:20 crc kubenswrapper[4959]: I1007 14:22:20.630742 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 14:22:20 crc kubenswrapper[4959]: I1007 14:22:20.632692 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d"} pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 14:22:20 crc kubenswrapper[4959]: I1007 14:22:20.632844 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" containerID="cri-o://0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d" gracePeriod=600 Oct 07 14:22:20 crc kubenswrapper[4959]: E1007 14:22:20.768215 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:22:21 crc kubenswrapper[4959]: I1007 14:22:21.343035 4959 generic.go:334] "Generic (PLEG): container finished" podID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerID="0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d" exitCode=0 Oct 07 14:22:21 crc kubenswrapper[4959]: I1007 14:22:21.343150 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerDied","Data":"0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d"} Oct 07 14:22:21 crc kubenswrapper[4959]: I1007 14:22:21.343204 4959 scope.go:117] "RemoveContainer" containerID="a084da1c14011d914fafa5905b78b9748fca6b5f32d398b4938e3e75516222b8" Oct 07 14:22:21 crc kubenswrapper[4959]: I1007 14:22:21.344230 4959 scope.go:117] "RemoveContainer" containerID="0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d" Oct 07 14:22:21 crc kubenswrapper[4959]: E1007 14:22:21.344740 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:22:35 crc kubenswrapper[4959]: I1007 14:22:35.654383 4959 scope.go:117] "RemoveContainer" containerID="0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d" Oct 07 14:22:35 crc kubenswrapper[4959]: E1007 14:22:35.657391 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:22:46 crc kubenswrapper[4959]: I1007 14:22:46.654076 4959 scope.go:117] "RemoveContainer" containerID="0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d" Oct 07 14:22:46 crc kubenswrapper[4959]: E1007 14:22:46.654873 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:22:58 crc kubenswrapper[4959]: I1007 14:22:58.661069 4959 scope.go:117] "RemoveContainer" containerID="0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d" Oct 07 14:22:58 crc kubenswrapper[4959]: E1007 14:22:58.661779 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:22:59 crc kubenswrapper[4959]: I1007 14:22:59.121816 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k"] Oct 07 14:22:59 crc kubenswrapper[4959]: I1007 14:22:59.129844 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6pz6k"] Oct 07 14:22:59 crc kubenswrapper[4959]: I1007 14:22:59.138985 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4"] Oct 07 14:22:59 crc kubenswrapper[4959]: I1007 14:22:59.152742 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-wjl8b"] Oct 07 14:22:59 crc kubenswrapper[4959]: I1007 14:22:59.159455 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkmr9"] Oct 07 14:22:59 crc kubenswrapper[4959]: I1007 14:22:59.165540 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5ctw4"] Oct 07 14:22:59 crc kubenswrapper[4959]: I1007 14:22:59.172417 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg"] Oct 07 14:22:59 crc kubenswrapper[4959]: I1007 14:22:59.179221 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q"] Oct 07 14:22:59 crc kubenswrapper[4959]: I1007 14:22:59.185868 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-jkmr9"] Oct 07 14:22:59 crc kubenswrapper[4959]: I1007 14:22:59.192590 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-km6rz"] Oct 07 14:22:59 crc kubenswrapper[4959]: I1007 14:22:59.198943 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc"] Oct 07 14:22:59 crc kubenswrapper[4959]: I1007 14:22:59.205640 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf"] Oct 07 14:22:59 crc kubenswrapper[4959]: I1007 14:22:59.212742 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-h5brv"] Oct 07 14:22:59 crc kubenswrapper[4959]: I1007 14:22:59.220643 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf"] Oct 07 14:22:59 crc kubenswrapper[4959]: I1007 14:22:59.226923 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-wjl8b"] Oct 07 14:22:59 crc kubenswrapper[4959]: I1007 14:22:59.233213 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7brtf"] Oct 07 14:22:59 crc kubenswrapper[4959]: I1007 14:22:59.239034 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-m9trg"] Oct 07 14:22:59 crc kubenswrapper[4959]: I1007 14:22:59.245224 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-h5brv"] Oct 07 14:22:59 crc kubenswrapper[4959]: I1007 14:22:59.253666 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vq5rc"] Oct 07 14:22:59 crc kubenswrapper[4959]: I1007 14:22:59.260864 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mdq6q"] Oct 07 14:22:59 crc kubenswrapper[4959]: I1007 14:22:59.269179 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-km6rz"] Oct 07 14:22:59 crc kubenswrapper[4959]: I1007 14:22:59.278546 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-m9xlf"] Oct 07 14:23:00 crc kubenswrapper[4959]: I1007 14:23:00.666748 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d310c8e-5f3a-484f-ab93-ee382d288df8" path="/var/lib/kubelet/pods/4d310c8e-5f3a-484f-ab93-ee382d288df8/volumes" Oct 07 14:23:00 crc kubenswrapper[4959]: I1007 14:23:00.667391 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52b87677-a4da-4477-9a8f-8a5a656ce4bc" path="/var/lib/kubelet/pods/52b87677-a4da-4477-9a8f-8a5a656ce4bc/volumes" Oct 07 14:23:00 crc kubenswrapper[4959]: I1007 14:23:00.667907 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88e3b980-0df4-43e6-bdf1-cab57e2732dc" path="/var/lib/kubelet/pods/88e3b980-0df4-43e6-bdf1-cab57e2732dc/volumes" Oct 07 14:23:00 crc kubenswrapper[4959]: I1007 14:23:00.668515 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac13d4c5-d59f-4a95-9253-8417785bc949" path="/var/lib/kubelet/pods/ac13d4c5-d59f-4a95-9253-8417785bc949/volumes" Oct 07 14:23:00 crc kubenswrapper[4959]: I1007 14:23:00.669567 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5887d2-a870-4086-b94b-4abdfefd42bd" path="/var/lib/kubelet/pods/bc5887d2-a870-4086-b94b-4abdfefd42bd/volumes" Oct 07 14:23:00 crc kubenswrapper[4959]: I1007 14:23:00.670076 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbc98733-35d4-477a-825f-3159319f68ad" path="/var/lib/kubelet/pods/cbc98733-35d4-477a-825f-3159319f68ad/volumes" Oct 07 14:23:00 crc kubenswrapper[4959]: I1007 14:23:00.670638 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d843db5d-75e6-4f95-a297-41af3451648f" path="/var/lib/kubelet/pods/d843db5d-75e6-4f95-a297-41af3451648f/volumes" Oct 07 14:23:00 crc kubenswrapper[4959]: I1007 14:23:00.671586 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d983b0f4-2357-46c9-bd51-32efc75ec771" path="/var/lib/kubelet/pods/d983b0f4-2357-46c9-bd51-32efc75ec771/volumes" Oct 07 14:23:00 crc kubenswrapper[4959]: I1007 14:23:00.672123 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2de54f5-c547-4a2c-afa3-2ba5715d423b" path="/var/lib/kubelet/pods/e2de54f5-c547-4a2c-afa3-2ba5715d423b/volumes" Oct 07 14:23:00 crc kubenswrapper[4959]: I1007 14:23:00.672598 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7967b3f-3a72-47d4-935d-4fa777b0fc46" path="/var/lib/kubelet/pods/e7967b3f-3a72-47d4-935d-4fa777b0fc46/volumes" Oct 07 14:23:00 crc kubenswrapper[4959]: I1007 14:23:00.673671 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fead8dac-8646-4828-be8e-d17bc43d7d61" path="/var/lib/kubelet/pods/fead8dac-8646-4828-be8e-d17bc43d7d61/volumes" Oct 07 14:23:04 crc kubenswrapper[4959]: I1007 14:23:04.987844 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v692z"] Oct 07 14:23:04 crc kubenswrapper[4959]: E1007 14:23:04.988911 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f98e86e-cc03-4344-82d6-897514556718" containerName="extract-utilities" Oct 07 14:23:04 crc kubenswrapper[4959]: I1007 14:23:04.988932 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f98e86e-cc03-4344-82d6-897514556718" containerName="extract-utilities" Oct 07 14:23:04 crc kubenswrapper[4959]: E1007 14:23:04.988965 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f98e86e-cc03-4344-82d6-897514556718" containerName="extract-content" Oct 07 14:23:04 crc kubenswrapper[4959]: I1007 14:23:04.988974 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f98e86e-cc03-4344-82d6-897514556718" containerName="extract-content" Oct 07 14:23:04 crc kubenswrapper[4959]: E1007 14:23:04.988993 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f98e86e-cc03-4344-82d6-897514556718" containerName="registry-server" Oct 07 14:23:04 crc kubenswrapper[4959]: I1007 14:23:04.989001 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f98e86e-cc03-4344-82d6-897514556718" containerName="registry-server" Oct 07 14:23:04 crc kubenswrapper[4959]: I1007 14:23:04.989233 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f98e86e-cc03-4344-82d6-897514556718" containerName="registry-server" Oct 07 14:23:04 crc kubenswrapper[4959]: I1007 14:23:04.990070 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v692z" Oct 07 14:23:04 crc kubenswrapper[4959]: I1007 14:23:04.992693 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 14:23:04 crc kubenswrapper[4959]: I1007 14:23:04.992864 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 14:23:04 crc kubenswrapper[4959]: I1007 14:23:04.992971 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-c6vmh" Oct 07 14:23:04 crc kubenswrapper[4959]: I1007 14:23:04.997938 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 14:23:04 crc kubenswrapper[4959]: I1007 14:23:04.997952 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 07 14:23:05 crc kubenswrapper[4959]: I1007 14:23:05.011712 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v692z"] Oct 07 14:23:05 crc kubenswrapper[4959]: I1007 14:23:05.103563 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4064e315-7d27-4e7a-b756-a5f01d7db24a-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-v692z\" (UID: \"4064e315-7d27-4e7a-b756-a5f01d7db24a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v692z" Oct 07 14:23:05 crc kubenswrapper[4959]: I1007 14:23:05.103613 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4064e315-7d27-4e7a-b756-a5f01d7db24a-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-v692z\" (UID: \"4064e315-7d27-4e7a-b756-a5f01d7db24a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v692z" Oct 07 14:23:05 crc kubenswrapper[4959]: I1007 14:23:05.103666 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4064e315-7d27-4e7a-b756-a5f01d7db24a-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-v692z\" (UID: \"4064e315-7d27-4e7a-b756-a5f01d7db24a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v692z" Oct 07 14:23:05 crc kubenswrapper[4959]: I1007 14:23:05.103719 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8tjs\" (UniqueName: \"kubernetes.io/projected/4064e315-7d27-4e7a-b756-a5f01d7db24a-kube-api-access-c8tjs\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-v692z\" (UID: \"4064e315-7d27-4e7a-b756-a5f01d7db24a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v692z" Oct 07 14:23:05 crc kubenswrapper[4959]: I1007 14:23:05.103830 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4064e315-7d27-4e7a-b756-a5f01d7db24a-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-v692z\" (UID: \"4064e315-7d27-4e7a-b756-a5f01d7db24a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v692z" Oct 07 14:23:05 crc kubenswrapper[4959]: I1007 14:23:05.205877 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4064e315-7d27-4e7a-b756-a5f01d7db24a-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-v692z\" (UID: \"4064e315-7d27-4e7a-b756-a5f01d7db24a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v692z" Oct 07 14:23:05 crc kubenswrapper[4959]: I1007 14:23:05.205957 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8tjs\" (UniqueName: \"kubernetes.io/projected/4064e315-7d27-4e7a-b756-a5f01d7db24a-kube-api-access-c8tjs\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-v692z\" (UID: \"4064e315-7d27-4e7a-b756-a5f01d7db24a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v692z" Oct 07 14:23:05 crc kubenswrapper[4959]: I1007 14:23:05.206076 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4064e315-7d27-4e7a-b756-a5f01d7db24a-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-v692z\" (UID: \"4064e315-7d27-4e7a-b756-a5f01d7db24a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v692z" Oct 07 14:23:05 crc kubenswrapper[4959]: I1007 14:23:05.206177 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4064e315-7d27-4e7a-b756-a5f01d7db24a-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-v692z\" (UID: \"4064e315-7d27-4e7a-b756-a5f01d7db24a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v692z" Oct 07 14:23:05 crc kubenswrapper[4959]: I1007 14:23:05.206205 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4064e315-7d27-4e7a-b756-a5f01d7db24a-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-v692z\" (UID: \"4064e315-7d27-4e7a-b756-a5f01d7db24a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v692z" Oct 07 14:23:05 crc kubenswrapper[4959]: I1007 14:23:05.211907 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4064e315-7d27-4e7a-b756-a5f01d7db24a-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-v692z\" (UID: \"4064e315-7d27-4e7a-b756-a5f01d7db24a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v692z" Oct 07 14:23:05 crc kubenswrapper[4959]: I1007 14:23:05.214680 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4064e315-7d27-4e7a-b756-a5f01d7db24a-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-v692z\" (UID: \"4064e315-7d27-4e7a-b756-a5f01d7db24a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v692z" Oct 07 14:23:05 crc kubenswrapper[4959]: I1007 14:23:05.215846 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4064e315-7d27-4e7a-b756-a5f01d7db24a-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-v692z\" (UID: \"4064e315-7d27-4e7a-b756-a5f01d7db24a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v692z" Oct 07 14:23:05 crc kubenswrapper[4959]: I1007 14:23:05.216017 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4064e315-7d27-4e7a-b756-a5f01d7db24a-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-v692z\" (UID: \"4064e315-7d27-4e7a-b756-a5f01d7db24a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v692z" Oct 07 14:23:05 crc kubenswrapper[4959]: I1007 14:23:05.233010 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8tjs\" (UniqueName: \"kubernetes.io/projected/4064e315-7d27-4e7a-b756-a5f01d7db24a-kube-api-access-c8tjs\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-v692z\" (UID: \"4064e315-7d27-4e7a-b756-a5f01d7db24a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v692z" Oct 07 14:23:05 crc kubenswrapper[4959]: I1007 14:23:05.309291 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v692z" Oct 07 14:23:05 crc kubenswrapper[4959]: I1007 14:23:05.856829 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v692z"] Oct 07 14:23:06 crc kubenswrapper[4959]: I1007 14:23:06.785033 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v692z" event={"ID":"4064e315-7d27-4e7a-b756-a5f01d7db24a","Type":"ContainerStarted","Data":"e63d767e38d679b8f020f8c6b679ef15420b60c088df923c0c75a23d2493cf1d"} Oct 07 14:23:06 crc kubenswrapper[4959]: I1007 14:23:06.785407 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v692z" event={"ID":"4064e315-7d27-4e7a-b756-a5f01d7db24a","Type":"ContainerStarted","Data":"21637d6fa81e63bd596f42e39ebf07c45b18fcd99716d4041ce646fec6ca58b3"} Oct 07 14:23:06 crc kubenswrapper[4959]: I1007 14:23:06.808128 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v692z" podStartSLOduration=2.31301783 podStartE2EDuration="2.808076759s" podCreationTimestamp="2025-10-07 14:23:04 +0000 UTC" firstStartedPulling="2025-10-07 14:23:05.865640393 +0000 UTC m=+2267.949044718" lastFinishedPulling="2025-10-07 14:23:06.360699322 +0000 UTC m=+2268.444103647" observedRunningTime="2025-10-07 14:23:06.800946053 +0000 UTC m=+2268.884350378" watchObservedRunningTime="2025-10-07 14:23:06.808076759 +0000 UTC m=+2268.891481104" Oct 07 14:23:13 crc kubenswrapper[4959]: I1007 14:23:13.653228 4959 scope.go:117] "RemoveContainer" containerID="0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d" Oct 07 14:23:13 crc kubenswrapper[4959]: E1007 14:23:13.653875 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:23:18 crc kubenswrapper[4959]: I1007 14:23:18.061640 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tvbfv"] Oct 07 14:23:18 crc kubenswrapper[4959]: I1007 14:23:18.064436 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tvbfv" Oct 07 14:23:18 crc kubenswrapper[4959]: I1007 14:23:18.079616 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tvbfv"] Oct 07 14:23:18 crc kubenswrapper[4959]: I1007 14:23:18.168665 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd11a0bc-88cf-47b0-b217-e61e24c82c4a-catalog-content\") pod \"community-operators-tvbfv\" (UID: \"bd11a0bc-88cf-47b0-b217-e61e24c82c4a\") " pod="openshift-marketplace/community-operators-tvbfv" Oct 07 14:23:18 crc kubenswrapper[4959]: I1007 14:23:18.168723 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd11a0bc-88cf-47b0-b217-e61e24c82c4a-utilities\") pod \"community-operators-tvbfv\" (UID: \"bd11a0bc-88cf-47b0-b217-e61e24c82c4a\") " pod="openshift-marketplace/community-operators-tvbfv" Oct 07 14:23:18 crc kubenswrapper[4959]: I1007 14:23:18.168757 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rr94w\" (UniqueName: \"kubernetes.io/projected/bd11a0bc-88cf-47b0-b217-e61e24c82c4a-kube-api-access-rr94w\") pod \"community-operators-tvbfv\" (UID: \"bd11a0bc-88cf-47b0-b217-e61e24c82c4a\") " pod="openshift-marketplace/community-operators-tvbfv" Oct 07 14:23:18 crc kubenswrapper[4959]: I1007 14:23:18.270956 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd11a0bc-88cf-47b0-b217-e61e24c82c4a-catalog-content\") pod \"community-operators-tvbfv\" (UID: \"bd11a0bc-88cf-47b0-b217-e61e24c82c4a\") " pod="openshift-marketplace/community-operators-tvbfv" Oct 07 14:23:18 crc kubenswrapper[4959]: I1007 14:23:18.271255 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd11a0bc-88cf-47b0-b217-e61e24c82c4a-utilities\") pod \"community-operators-tvbfv\" (UID: \"bd11a0bc-88cf-47b0-b217-e61e24c82c4a\") " pod="openshift-marketplace/community-operators-tvbfv" Oct 07 14:23:18 crc kubenswrapper[4959]: I1007 14:23:18.271360 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rr94w\" (UniqueName: \"kubernetes.io/projected/bd11a0bc-88cf-47b0-b217-e61e24c82c4a-kube-api-access-rr94w\") pod \"community-operators-tvbfv\" (UID: \"bd11a0bc-88cf-47b0-b217-e61e24c82c4a\") " pod="openshift-marketplace/community-operators-tvbfv" Oct 07 14:23:18 crc kubenswrapper[4959]: I1007 14:23:18.271563 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd11a0bc-88cf-47b0-b217-e61e24c82c4a-catalog-content\") pod \"community-operators-tvbfv\" (UID: \"bd11a0bc-88cf-47b0-b217-e61e24c82c4a\") " pod="openshift-marketplace/community-operators-tvbfv" Oct 07 14:23:18 crc kubenswrapper[4959]: I1007 14:23:18.271741 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd11a0bc-88cf-47b0-b217-e61e24c82c4a-utilities\") pod \"community-operators-tvbfv\" (UID: \"bd11a0bc-88cf-47b0-b217-e61e24c82c4a\") " pod="openshift-marketplace/community-operators-tvbfv" Oct 07 14:23:18 crc kubenswrapper[4959]: I1007 14:23:18.291390 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rr94w\" (UniqueName: \"kubernetes.io/projected/bd11a0bc-88cf-47b0-b217-e61e24c82c4a-kube-api-access-rr94w\") pod \"community-operators-tvbfv\" (UID: \"bd11a0bc-88cf-47b0-b217-e61e24c82c4a\") " pod="openshift-marketplace/community-operators-tvbfv" Oct 07 14:23:18 crc kubenswrapper[4959]: I1007 14:23:18.385280 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tvbfv" Oct 07 14:23:18 crc kubenswrapper[4959]: I1007 14:23:18.906028 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tvbfv"] Oct 07 14:23:19 crc kubenswrapper[4959]: I1007 14:23:19.912775 4959 generic.go:334] "Generic (PLEG): container finished" podID="bd11a0bc-88cf-47b0-b217-e61e24c82c4a" containerID="d689bcbb0af79183954953a564fb0e391290dfc49a77d304b857a58531495afc" exitCode=0 Oct 07 14:23:19 crc kubenswrapper[4959]: I1007 14:23:19.912909 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvbfv" event={"ID":"bd11a0bc-88cf-47b0-b217-e61e24c82c4a","Type":"ContainerDied","Data":"d689bcbb0af79183954953a564fb0e391290dfc49a77d304b857a58531495afc"} Oct 07 14:23:19 crc kubenswrapper[4959]: I1007 14:23:19.913266 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvbfv" event={"ID":"bd11a0bc-88cf-47b0-b217-e61e24c82c4a","Type":"ContainerStarted","Data":"3aabcbe3d3dbaeca1373c64f89ffb9a75c483b176fce1bd8a7c7259031a2c28b"} Oct 07 14:23:19 crc kubenswrapper[4959]: I1007 14:23:19.917572 4959 generic.go:334] "Generic (PLEG): container finished" podID="4064e315-7d27-4e7a-b756-a5f01d7db24a" containerID="e63d767e38d679b8f020f8c6b679ef15420b60c088df923c0c75a23d2493cf1d" exitCode=0 Oct 07 14:23:19 crc kubenswrapper[4959]: I1007 14:23:19.917607 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v692z" event={"ID":"4064e315-7d27-4e7a-b756-a5f01d7db24a","Type":"ContainerDied","Data":"e63d767e38d679b8f020f8c6b679ef15420b60c088df923c0c75a23d2493cf1d"} Oct 07 14:23:20 crc kubenswrapper[4959]: I1007 14:23:20.929032 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvbfv" event={"ID":"bd11a0bc-88cf-47b0-b217-e61e24c82c4a","Type":"ContainerStarted","Data":"7ea640db6f84487bda29d1428d75cb54e4a5908b9c9d74ee5f4038d7777131bf"} Oct 07 14:23:21 crc kubenswrapper[4959]: I1007 14:23:21.399465 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v692z" Oct 07 14:23:21 crc kubenswrapper[4959]: I1007 14:23:21.562772 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4064e315-7d27-4e7a-b756-a5f01d7db24a-repo-setup-combined-ca-bundle\") pod \"4064e315-7d27-4e7a-b756-a5f01d7db24a\" (UID: \"4064e315-7d27-4e7a-b756-a5f01d7db24a\") " Oct 07 14:23:21 crc kubenswrapper[4959]: I1007 14:23:21.562977 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4064e315-7d27-4e7a-b756-a5f01d7db24a-inventory\") pod \"4064e315-7d27-4e7a-b756-a5f01d7db24a\" (UID: \"4064e315-7d27-4e7a-b756-a5f01d7db24a\") " Oct 07 14:23:21 crc kubenswrapper[4959]: I1007 14:23:21.563085 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8tjs\" (UniqueName: \"kubernetes.io/projected/4064e315-7d27-4e7a-b756-a5f01d7db24a-kube-api-access-c8tjs\") pod \"4064e315-7d27-4e7a-b756-a5f01d7db24a\" (UID: \"4064e315-7d27-4e7a-b756-a5f01d7db24a\") " Oct 07 14:23:21 crc kubenswrapper[4959]: I1007 14:23:21.563147 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4064e315-7d27-4e7a-b756-a5f01d7db24a-ceph\") pod \"4064e315-7d27-4e7a-b756-a5f01d7db24a\" (UID: \"4064e315-7d27-4e7a-b756-a5f01d7db24a\") " Oct 07 14:23:21 crc kubenswrapper[4959]: I1007 14:23:21.563330 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4064e315-7d27-4e7a-b756-a5f01d7db24a-ssh-key\") pod \"4064e315-7d27-4e7a-b756-a5f01d7db24a\" (UID: \"4064e315-7d27-4e7a-b756-a5f01d7db24a\") " Oct 07 14:23:21 crc kubenswrapper[4959]: I1007 14:23:21.570581 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4064e315-7d27-4e7a-b756-a5f01d7db24a-kube-api-access-c8tjs" (OuterVolumeSpecName: "kube-api-access-c8tjs") pod "4064e315-7d27-4e7a-b756-a5f01d7db24a" (UID: "4064e315-7d27-4e7a-b756-a5f01d7db24a"). InnerVolumeSpecName "kube-api-access-c8tjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:23:21 crc kubenswrapper[4959]: I1007 14:23:21.571282 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4064e315-7d27-4e7a-b756-a5f01d7db24a-ceph" (OuterVolumeSpecName: "ceph") pod "4064e315-7d27-4e7a-b756-a5f01d7db24a" (UID: "4064e315-7d27-4e7a-b756-a5f01d7db24a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:23:21 crc kubenswrapper[4959]: I1007 14:23:21.579281 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4064e315-7d27-4e7a-b756-a5f01d7db24a-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "4064e315-7d27-4e7a-b756-a5f01d7db24a" (UID: "4064e315-7d27-4e7a-b756-a5f01d7db24a"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:23:21 crc kubenswrapper[4959]: I1007 14:23:21.590593 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4064e315-7d27-4e7a-b756-a5f01d7db24a-inventory" (OuterVolumeSpecName: "inventory") pod "4064e315-7d27-4e7a-b756-a5f01d7db24a" (UID: "4064e315-7d27-4e7a-b756-a5f01d7db24a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:23:21 crc kubenswrapper[4959]: I1007 14:23:21.597054 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4064e315-7d27-4e7a-b756-a5f01d7db24a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4064e315-7d27-4e7a-b756-a5f01d7db24a" (UID: "4064e315-7d27-4e7a-b756-a5f01d7db24a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:23:21 crc kubenswrapper[4959]: I1007 14:23:21.666727 4959 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4064e315-7d27-4e7a-b756-a5f01d7db24a-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:23:21 crc kubenswrapper[4959]: I1007 14:23:21.666791 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4064e315-7d27-4e7a-b756-a5f01d7db24a-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 14:23:21 crc kubenswrapper[4959]: I1007 14:23:21.666812 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8tjs\" (UniqueName: \"kubernetes.io/projected/4064e315-7d27-4e7a-b756-a5f01d7db24a-kube-api-access-c8tjs\") on node \"crc\" DevicePath \"\"" Oct 07 14:23:21 crc kubenswrapper[4959]: I1007 14:23:21.666837 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4064e315-7d27-4e7a-b756-a5f01d7db24a-ceph\") on node \"crc\" DevicePath \"\"" Oct 07 14:23:21 crc kubenswrapper[4959]: I1007 14:23:21.666855 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4064e315-7d27-4e7a-b756-a5f01d7db24a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 14:23:21 crc kubenswrapper[4959]: I1007 14:23:21.941367 4959 generic.go:334] "Generic (PLEG): container finished" podID="bd11a0bc-88cf-47b0-b217-e61e24c82c4a" containerID="7ea640db6f84487bda29d1428d75cb54e4a5908b9c9d74ee5f4038d7777131bf" exitCode=0 Oct 07 14:23:21 crc kubenswrapper[4959]: I1007 14:23:21.941502 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvbfv" event={"ID":"bd11a0bc-88cf-47b0-b217-e61e24c82c4a","Type":"ContainerDied","Data":"7ea640db6f84487bda29d1428d75cb54e4a5908b9c9d74ee5f4038d7777131bf"} Oct 07 14:23:21 crc kubenswrapper[4959]: I1007 14:23:21.943749 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v692z" event={"ID":"4064e315-7d27-4e7a-b756-a5f01d7db24a","Type":"ContainerDied","Data":"21637d6fa81e63bd596f42e39ebf07c45b18fcd99716d4041ce646fec6ca58b3"} Oct 07 14:23:21 crc kubenswrapper[4959]: I1007 14:23:21.943785 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21637d6fa81e63bd596f42e39ebf07c45b18fcd99716d4041ce646fec6ca58b3" Oct 07 14:23:21 crc kubenswrapper[4959]: I1007 14:23:21.943891 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v692z" Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.034045 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl"] Oct 07 14:23:22 crc kubenswrapper[4959]: E1007 14:23:22.034989 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4064e315-7d27-4e7a-b756-a5f01d7db24a" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.035032 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4064e315-7d27-4e7a-b756-a5f01d7db24a" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.035490 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="4064e315-7d27-4e7a-b756-a5f01d7db24a" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.037512 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl" Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.040696 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.040816 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.040850 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.041170 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.042744 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-c6vmh" Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.044284 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl"] Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.179831 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ddf979d-a363-420d-8368-aac337f5b078-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl\" (UID: \"4ddf979d-a363-420d-8368-aac337f5b078\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl" Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.180398 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4ddf979d-a363-420d-8368-aac337f5b078-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl\" (UID: \"4ddf979d-a363-420d-8368-aac337f5b078\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl" Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.180523 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ddf979d-a363-420d-8368-aac337f5b078-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl\" (UID: \"4ddf979d-a363-420d-8368-aac337f5b078\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl" Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.180793 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlw5s\" (UniqueName: \"kubernetes.io/projected/4ddf979d-a363-420d-8368-aac337f5b078-kube-api-access-rlw5s\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl\" (UID: \"4ddf979d-a363-420d-8368-aac337f5b078\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl" Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.181012 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ddf979d-a363-420d-8368-aac337f5b078-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl\" (UID: \"4ddf979d-a363-420d-8368-aac337f5b078\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl" Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.282383 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ddf979d-a363-420d-8368-aac337f5b078-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl\" (UID: \"4ddf979d-a363-420d-8368-aac337f5b078\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl" Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.282701 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ddf979d-a363-420d-8368-aac337f5b078-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl\" (UID: \"4ddf979d-a363-420d-8368-aac337f5b078\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl" Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.282837 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4ddf979d-a363-420d-8368-aac337f5b078-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl\" (UID: \"4ddf979d-a363-420d-8368-aac337f5b078\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl" Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.282927 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ddf979d-a363-420d-8368-aac337f5b078-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl\" (UID: \"4ddf979d-a363-420d-8368-aac337f5b078\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl" Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.283010 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlw5s\" (UniqueName: \"kubernetes.io/projected/4ddf979d-a363-420d-8368-aac337f5b078-kube-api-access-rlw5s\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl\" (UID: \"4ddf979d-a363-420d-8368-aac337f5b078\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl" Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.286674 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ddf979d-a363-420d-8368-aac337f5b078-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl\" (UID: \"4ddf979d-a363-420d-8368-aac337f5b078\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl" Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.294398 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ddf979d-a363-420d-8368-aac337f5b078-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl\" (UID: \"4ddf979d-a363-420d-8368-aac337f5b078\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl" Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.294754 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ddf979d-a363-420d-8368-aac337f5b078-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl\" (UID: \"4ddf979d-a363-420d-8368-aac337f5b078\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl" Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.295713 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4ddf979d-a363-420d-8368-aac337f5b078-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl\" (UID: \"4ddf979d-a363-420d-8368-aac337f5b078\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl" Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.303737 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlw5s\" (UniqueName: \"kubernetes.io/projected/4ddf979d-a363-420d-8368-aac337f5b078-kube-api-access-rlw5s\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl\" (UID: \"4ddf979d-a363-420d-8368-aac337f5b078\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl" Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.367310 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl" Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.924937 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl"] Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.952648 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl" event={"ID":"4ddf979d-a363-420d-8368-aac337f5b078","Type":"ContainerStarted","Data":"d9d91c65daad29d3975bf660f797f7436057a704c5c1568dd04bd9b0b6ea72a1"} Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.969062 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvbfv" event={"ID":"bd11a0bc-88cf-47b0-b217-e61e24c82c4a","Type":"ContainerStarted","Data":"d3fb28baea217a843b6e8991b68932d5f8127d18d21c65cc016734da1be3ac82"} Oct 07 14:23:22 crc kubenswrapper[4959]: I1007 14:23:22.991849 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tvbfv" podStartSLOduration=2.348440521 podStartE2EDuration="4.991829347s" podCreationTimestamp="2025-10-07 14:23:18 +0000 UTC" firstStartedPulling="2025-10-07 14:23:19.916880694 +0000 UTC m=+2282.000285019" lastFinishedPulling="2025-10-07 14:23:22.56026952 +0000 UTC m=+2284.643673845" observedRunningTime="2025-10-07 14:23:22.986508925 +0000 UTC m=+2285.069913260" watchObservedRunningTime="2025-10-07 14:23:22.991829347 +0000 UTC m=+2285.075233672" Oct 07 14:23:23 crc kubenswrapper[4959]: I1007 14:23:23.978749 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl" event={"ID":"4ddf979d-a363-420d-8368-aac337f5b078","Type":"ContainerStarted","Data":"ee59181864353ea5d528ed992d635681ac34b630c431cdbbcf8461856152589d"} Oct 07 14:23:23 crc kubenswrapper[4959]: I1007 14:23:23.998144 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl" podStartSLOduration=1.506151992 podStartE2EDuration="1.998123865s" podCreationTimestamp="2025-10-07 14:23:22 +0000 UTC" firstStartedPulling="2025-10-07 14:23:22.926879278 +0000 UTC m=+2285.010283603" lastFinishedPulling="2025-10-07 14:23:23.418851151 +0000 UTC m=+2285.502255476" observedRunningTime="2025-10-07 14:23:23.991791138 +0000 UTC m=+2286.075195463" watchObservedRunningTime="2025-10-07 14:23:23.998123865 +0000 UTC m=+2286.081528190" Oct 07 14:23:26 crc kubenswrapper[4959]: I1007 14:23:26.655216 4959 scope.go:117] "RemoveContainer" containerID="0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d" Oct 07 14:23:26 crc kubenswrapper[4959]: E1007 14:23:26.655914 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:23:28 crc kubenswrapper[4959]: I1007 14:23:28.385820 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tvbfv" Oct 07 14:23:28 crc kubenswrapper[4959]: I1007 14:23:28.386203 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tvbfv" Oct 07 14:23:28 crc kubenswrapper[4959]: I1007 14:23:28.433779 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tvbfv" Oct 07 14:23:29 crc kubenswrapper[4959]: I1007 14:23:29.082772 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tvbfv" Oct 07 14:23:29 crc kubenswrapper[4959]: I1007 14:23:29.126585 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tvbfv"] Oct 07 14:23:29 crc kubenswrapper[4959]: I1007 14:23:29.600555 4959 scope.go:117] "RemoveContainer" containerID="18bd1574eb0bda3cf5dd8824be4eba6c0e4e2153c966f6cd54609346fb1a8a09" Oct 07 14:23:29 crc kubenswrapper[4959]: I1007 14:23:29.635412 4959 scope.go:117] "RemoveContainer" containerID="aefc8d2781df2b525eacaaa3711fe17975d192444d61361f987caf6586ee2a0d" Oct 07 14:23:29 crc kubenswrapper[4959]: I1007 14:23:29.692989 4959 scope.go:117] "RemoveContainer" containerID="35091a54cbb649c1df69456b4413f744e19f37ada482440213a5881c959b6079" Oct 07 14:23:29 crc kubenswrapper[4959]: I1007 14:23:29.763648 4959 scope.go:117] "RemoveContainer" containerID="eb5fe6d5b2fa51a8cbb00614d1816c41eb1ca032e53c3477ddfa2f2a340b2a0a" Oct 07 14:23:29 crc kubenswrapper[4959]: I1007 14:23:29.796090 4959 scope.go:117] "RemoveContainer" containerID="412050b987b86adccfa498c30b05e722afed7366b18779c3715eeae0cfd661a8" Oct 07 14:23:29 crc kubenswrapper[4959]: I1007 14:23:29.837272 4959 scope.go:117] "RemoveContainer" containerID="1c5f573eb9c4a3ab7894dbab43f9e66531c2be377ebf98a67796c9c94b35f235" Oct 07 14:23:29 crc kubenswrapper[4959]: I1007 14:23:29.894187 4959 scope.go:117] "RemoveContainer" containerID="d595881cf5fc11dbf8725f0a0c36208db33f3dd42dc9d45d72caef1faf92d118" Oct 07 14:23:29 crc kubenswrapper[4959]: I1007 14:23:29.920770 4959 scope.go:117] "RemoveContainer" containerID="f2f5e667a2b6d76a4b24d950849f384c46ece4c029f8e6fdd40a681134a2bc00" Oct 07 14:23:29 crc kubenswrapper[4959]: I1007 14:23:29.957447 4959 scope.go:117] "RemoveContainer" containerID="fdbec70c8246df00d7cf23865b4a11be9cf1cfbcfd29e732d8e4d539b6c8bb0e" Oct 07 14:23:29 crc kubenswrapper[4959]: I1007 14:23:29.980814 4959 scope.go:117] "RemoveContainer" containerID="6540e28dc68f53ae9ebcd09d0b1dc6198da7c5ed82f03c4807c4af7669f43176" Oct 07 14:23:30 crc kubenswrapper[4959]: I1007 14:23:30.021959 4959 scope.go:117] "RemoveContainer" containerID="8c2b3d835cf81c27fb72a2add3653dfa9fd1caf6bb28054442b2e29a6e0f8742" Oct 07 14:23:31 crc kubenswrapper[4959]: I1007 14:23:31.048764 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tvbfv" podUID="bd11a0bc-88cf-47b0-b217-e61e24c82c4a" containerName="registry-server" containerID="cri-o://d3fb28baea217a843b6e8991b68932d5f8127d18d21c65cc016734da1be3ac82" gracePeriod=2 Oct 07 14:23:31 crc kubenswrapper[4959]: I1007 14:23:31.490197 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tvbfv" Oct 07 14:23:31 crc kubenswrapper[4959]: I1007 14:23:31.596235 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd11a0bc-88cf-47b0-b217-e61e24c82c4a-catalog-content\") pod \"bd11a0bc-88cf-47b0-b217-e61e24c82c4a\" (UID: \"bd11a0bc-88cf-47b0-b217-e61e24c82c4a\") " Oct 07 14:23:31 crc kubenswrapper[4959]: I1007 14:23:31.596328 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rr94w\" (UniqueName: \"kubernetes.io/projected/bd11a0bc-88cf-47b0-b217-e61e24c82c4a-kube-api-access-rr94w\") pod \"bd11a0bc-88cf-47b0-b217-e61e24c82c4a\" (UID: \"bd11a0bc-88cf-47b0-b217-e61e24c82c4a\") " Oct 07 14:23:31 crc kubenswrapper[4959]: I1007 14:23:31.596745 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd11a0bc-88cf-47b0-b217-e61e24c82c4a-utilities\") pod \"bd11a0bc-88cf-47b0-b217-e61e24c82c4a\" (UID: \"bd11a0bc-88cf-47b0-b217-e61e24c82c4a\") " Oct 07 14:23:31 crc kubenswrapper[4959]: I1007 14:23:31.597593 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd11a0bc-88cf-47b0-b217-e61e24c82c4a-utilities" (OuterVolumeSpecName: "utilities") pod "bd11a0bc-88cf-47b0-b217-e61e24c82c4a" (UID: "bd11a0bc-88cf-47b0-b217-e61e24c82c4a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:23:31 crc kubenswrapper[4959]: I1007 14:23:31.601311 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd11a0bc-88cf-47b0-b217-e61e24c82c4a-kube-api-access-rr94w" (OuterVolumeSpecName: "kube-api-access-rr94w") pod "bd11a0bc-88cf-47b0-b217-e61e24c82c4a" (UID: "bd11a0bc-88cf-47b0-b217-e61e24c82c4a"). InnerVolumeSpecName "kube-api-access-rr94w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:23:31 crc kubenswrapper[4959]: I1007 14:23:31.699358 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd11a0bc-88cf-47b0-b217-e61e24c82c4a-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 14:23:31 crc kubenswrapper[4959]: I1007 14:23:31.699390 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rr94w\" (UniqueName: \"kubernetes.io/projected/bd11a0bc-88cf-47b0-b217-e61e24c82c4a-kube-api-access-rr94w\") on node \"crc\" DevicePath \"\"" Oct 07 14:23:31 crc kubenswrapper[4959]: I1007 14:23:31.963749 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd11a0bc-88cf-47b0-b217-e61e24c82c4a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bd11a0bc-88cf-47b0-b217-e61e24c82c4a" (UID: "bd11a0bc-88cf-47b0-b217-e61e24c82c4a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:23:32 crc kubenswrapper[4959]: I1007 14:23:32.005437 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd11a0bc-88cf-47b0-b217-e61e24c82c4a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 14:23:32 crc kubenswrapper[4959]: I1007 14:23:32.063423 4959 generic.go:334] "Generic (PLEG): container finished" podID="bd11a0bc-88cf-47b0-b217-e61e24c82c4a" containerID="d3fb28baea217a843b6e8991b68932d5f8127d18d21c65cc016734da1be3ac82" exitCode=0 Oct 07 14:23:32 crc kubenswrapper[4959]: I1007 14:23:32.063477 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvbfv" event={"ID":"bd11a0bc-88cf-47b0-b217-e61e24c82c4a","Type":"ContainerDied","Data":"d3fb28baea217a843b6e8991b68932d5f8127d18d21c65cc016734da1be3ac82"} Oct 07 14:23:32 crc kubenswrapper[4959]: I1007 14:23:32.063531 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvbfv" event={"ID":"bd11a0bc-88cf-47b0-b217-e61e24c82c4a","Type":"ContainerDied","Data":"3aabcbe3d3dbaeca1373c64f89ffb9a75c483b176fce1bd8a7c7259031a2c28b"} Oct 07 14:23:32 crc kubenswrapper[4959]: I1007 14:23:32.063553 4959 scope.go:117] "RemoveContainer" containerID="d3fb28baea217a843b6e8991b68932d5f8127d18d21c65cc016734da1be3ac82" Oct 07 14:23:32 crc kubenswrapper[4959]: I1007 14:23:32.063781 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tvbfv" Oct 07 14:23:32 crc kubenswrapper[4959]: I1007 14:23:32.096364 4959 scope.go:117] "RemoveContainer" containerID="7ea640db6f84487bda29d1428d75cb54e4a5908b9c9d74ee5f4038d7777131bf" Oct 07 14:23:32 crc kubenswrapper[4959]: I1007 14:23:32.126747 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tvbfv"] Oct 07 14:23:32 crc kubenswrapper[4959]: I1007 14:23:32.131807 4959 scope.go:117] "RemoveContainer" containerID="d689bcbb0af79183954953a564fb0e391290dfc49a77d304b857a58531495afc" Oct 07 14:23:32 crc kubenswrapper[4959]: I1007 14:23:32.133798 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tvbfv"] Oct 07 14:23:32 crc kubenswrapper[4959]: I1007 14:23:32.164639 4959 scope.go:117] "RemoveContainer" containerID="d3fb28baea217a843b6e8991b68932d5f8127d18d21c65cc016734da1be3ac82" Oct 07 14:23:32 crc kubenswrapper[4959]: E1007 14:23:32.165605 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3fb28baea217a843b6e8991b68932d5f8127d18d21c65cc016734da1be3ac82\": container with ID starting with d3fb28baea217a843b6e8991b68932d5f8127d18d21c65cc016734da1be3ac82 not found: ID does not exist" containerID="d3fb28baea217a843b6e8991b68932d5f8127d18d21c65cc016734da1be3ac82" Oct 07 14:23:32 crc kubenswrapper[4959]: I1007 14:23:32.165637 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3fb28baea217a843b6e8991b68932d5f8127d18d21c65cc016734da1be3ac82"} err="failed to get container status \"d3fb28baea217a843b6e8991b68932d5f8127d18d21c65cc016734da1be3ac82\": rpc error: code = NotFound desc = could not find container \"d3fb28baea217a843b6e8991b68932d5f8127d18d21c65cc016734da1be3ac82\": container with ID starting with d3fb28baea217a843b6e8991b68932d5f8127d18d21c65cc016734da1be3ac82 not found: ID does not exist" Oct 07 14:23:32 crc kubenswrapper[4959]: I1007 14:23:32.165658 4959 scope.go:117] "RemoveContainer" containerID="7ea640db6f84487bda29d1428d75cb54e4a5908b9c9d74ee5f4038d7777131bf" Oct 07 14:23:32 crc kubenswrapper[4959]: E1007 14:23:32.166371 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ea640db6f84487bda29d1428d75cb54e4a5908b9c9d74ee5f4038d7777131bf\": container with ID starting with 7ea640db6f84487bda29d1428d75cb54e4a5908b9c9d74ee5f4038d7777131bf not found: ID does not exist" containerID="7ea640db6f84487bda29d1428d75cb54e4a5908b9c9d74ee5f4038d7777131bf" Oct 07 14:23:32 crc kubenswrapper[4959]: I1007 14:23:32.166438 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ea640db6f84487bda29d1428d75cb54e4a5908b9c9d74ee5f4038d7777131bf"} err="failed to get container status \"7ea640db6f84487bda29d1428d75cb54e4a5908b9c9d74ee5f4038d7777131bf\": rpc error: code = NotFound desc = could not find container \"7ea640db6f84487bda29d1428d75cb54e4a5908b9c9d74ee5f4038d7777131bf\": container with ID starting with 7ea640db6f84487bda29d1428d75cb54e4a5908b9c9d74ee5f4038d7777131bf not found: ID does not exist" Oct 07 14:23:32 crc kubenswrapper[4959]: I1007 14:23:32.166488 4959 scope.go:117] "RemoveContainer" containerID="d689bcbb0af79183954953a564fb0e391290dfc49a77d304b857a58531495afc" Oct 07 14:23:32 crc kubenswrapper[4959]: E1007 14:23:32.166898 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d689bcbb0af79183954953a564fb0e391290dfc49a77d304b857a58531495afc\": container with ID starting with d689bcbb0af79183954953a564fb0e391290dfc49a77d304b857a58531495afc not found: ID does not exist" containerID="d689bcbb0af79183954953a564fb0e391290dfc49a77d304b857a58531495afc" Oct 07 14:23:32 crc kubenswrapper[4959]: I1007 14:23:32.166927 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d689bcbb0af79183954953a564fb0e391290dfc49a77d304b857a58531495afc"} err="failed to get container status \"d689bcbb0af79183954953a564fb0e391290dfc49a77d304b857a58531495afc\": rpc error: code = NotFound desc = could not find container \"d689bcbb0af79183954953a564fb0e391290dfc49a77d304b857a58531495afc\": container with ID starting with d689bcbb0af79183954953a564fb0e391290dfc49a77d304b857a58531495afc not found: ID does not exist" Oct 07 14:23:32 crc kubenswrapper[4959]: I1007 14:23:32.664527 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd11a0bc-88cf-47b0-b217-e61e24c82c4a" path="/var/lib/kubelet/pods/bd11a0bc-88cf-47b0-b217-e61e24c82c4a/volumes" Oct 07 14:23:37 crc kubenswrapper[4959]: I1007 14:23:37.654189 4959 scope.go:117] "RemoveContainer" containerID="0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d" Oct 07 14:23:37 crc kubenswrapper[4959]: E1007 14:23:37.655544 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:23:52 crc kubenswrapper[4959]: I1007 14:23:52.653427 4959 scope.go:117] "RemoveContainer" containerID="0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d" Oct 07 14:23:52 crc kubenswrapper[4959]: E1007 14:23:52.654302 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:24:06 crc kubenswrapper[4959]: I1007 14:24:06.653880 4959 scope.go:117] "RemoveContainer" containerID="0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d" Oct 07 14:24:06 crc kubenswrapper[4959]: E1007 14:24:06.655011 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:24:19 crc kubenswrapper[4959]: I1007 14:24:19.653773 4959 scope.go:117] "RemoveContainer" containerID="0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d" Oct 07 14:24:19 crc kubenswrapper[4959]: E1007 14:24:19.654707 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:24:34 crc kubenswrapper[4959]: I1007 14:24:34.654051 4959 scope.go:117] "RemoveContainer" containerID="0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d" Oct 07 14:24:34 crc kubenswrapper[4959]: E1007 14:24:34.655005 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:24:47 crc kubenswrapper[4959]: I1007 14:24:47.653485 4959 scope.go:117] "RemoveContainer" containerID="0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d" Oct 07 14:24:47 crc kubenswrapper[4959]: E1007 14:24:47.654353 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:24:59 crc kubenswrapper[4959]: I1007 14:24:59.833258 4959 generic.go:334] "Generic (PLEG): container finished" podID="4ddf979d-a363-420d-8368-aac337f5b078" containerID="ee59181864353ea5d528ed992d635681ac34b630c431cdbbcf8461856152589d" exitCode=0 Oct 07 14:24:59 crc kubenswrapper[4959]: I1007 14:24:59.833353 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl" event={"ID":"4ddf979d-a363-420d-8368-aac337f5b078","Type":"ContainerDied","Data":"ee59181864353ea5d528ed992d635681ac34b630c431cdbbcf8461856152589d"} Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.291688 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl" Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.421849 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlw5s\" (UniqueName: \"kubernetes.io/projected/4ddf979d-a363-420d-8368-aac337f5b078-kube-api-access-rlw5s\") pod \"4ddf979d-a363-420d-8368-aac337f5b078\" (UID: \"4ddf979d-a363-420d-8368-aac337f5b078\") " Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.421980 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ddf979d-a363-420d-8368-aac337f5b078-ssh-key\") pod \"4ddf979d-a363-420d-8368-aac337f5b078\" (UID: \"4ddf979d-a363-420d-8368-aac337f5b078\") " Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.423179 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ddf979d-a363-420d-8368-aac337f5b078-bootstrap-combined-ca-bundle\") pod \"4ddf979d-a363-420d-8368-aac337f5b078\" (UID: \"4ddf979d-a363-420d-8368-aac337f5b078\") " Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.423348 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ddf979d-a363-420d-8368-aac337f5b078-inventory\") pod \"4ddf979d-a363-420d-8368-aac337f5b078\" (UID: \"4ddf979d-a363-420d-8368-aac337f5b078\") " Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.423488 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4ddf979d-a363-420d-8368-aac337f5b078-ceph\") pod \"4ddf979d-a363-420d-8368-aac337f5b078\" (UID: \"4ddf979d-a363-420d-8368-aac337f5b078\") " Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.430615 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ddf979d-a363-420d-8368-aac337f5b078-kube-api-access-rlw5s" (OuterVolumeSpecName: "kube-api-access-rlw5s") pod "4ddf979d-a363-420d-8368-aac337f5b078" (UID: "4ddf979d-a363-420d-8368-aac337f5b078"). InnerVolumeSpecName "kube-api-access-rlw5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.431663 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ddf979d-a363-420d-8368-aac337f5b078-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "4ddf979d-a363-420d-8368-aac337f5b078" (UID: "4ddf979d-a363-420d-8368-aac337f5b078"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.431844 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ddf979d-a363-420d-8368-aac337f5b078-ceph" (OuterVolumeSpecName: "ceph") pod "4ddf979d-a363-420d-8368-aac337f5b078" (UID: "4ddf979d-a363-420d-8368-aac337f5b078"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.454475 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ddf979d-a363-420d-8368-aac337f5b078-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4ddf979d-a363-420d-8368-aac337f5b078" (UID: "4ddf979d-a363-420d-8368-aac337f5b078"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.464143 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ddf979d-a363-420d-8368-aac337f5b078-inventory" (OuterVolumeSpecName: "inventory") pod "4ddf979d-a363-420d-8368-aac337f5b078" (UID: "4ddf979d-a363-420d-8368-aac337f5b078"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.526183 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ddf979d-a363-420d-8368-aac337f5b078-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.526364 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4ddf979d-a363-420d-8368-aac337f5b078-ceph\") on node \"crc\" DevicePath \"\"" Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.526453 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlw5s\" (UniqueName: \"kubernetes.io/projected/4ddf979d-a363-420d-8368-aac337f5b078-kube-api-access-rlw5s\") on node \"crc\" DevicePath \"\"" Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.526536 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ddf979d-a363-420d-8368-aac337f5b078-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.526618 4959 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ddf979d-a363-420d-8368-aac337f5b078-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.855614 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl" event={"ID":"4ddf979d-a363-420d-8368-aac337f5b078","Type":"ContainerDied","Data":"d9d91c65daad29d3975bf660f797f7436057a704c5c1568dd04bd9b0b6ea72a1"} Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.855696 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9d91c65daad29d3975bf660f797f7436057a704c5c1568dd04bd9b0b6ea72a1" Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.855887 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl" Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.954341 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-khd49"] Oct 07 14:25:01 crc kubenswrapper[4959]: E1007 14:25:01.954836 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd11a0bc-88cf-47b0-b217-e61e24c82c4a" containerName="extract-content" Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.954859 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd11a0bc-88cf-47b0-b217-e61e24c82c4a" containerName="extract-content" Oct 07 14:25:01 crc kubenswrapper[4959]: E1007 14:25:01.954887 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ddf979d-a363-420d-8368-aac337f5b078" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.954897 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ddf979d-a363-420d-8368-aac337f5b078" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 07 14:25:01 crc kubenswrapper[4959]: E1007 14:25:01.954931 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd11a0bc-88cf-47b0-b217-e61e24c82c4a" containerName="extract-utilities" Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.954938 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd11a0bc-88cf-47b0-b217-e61e24c82c4a" containerName="extract-utilities" Oct 07 14:25:01 crc kubenswrapper[4959]: E1007 14:25:01.954949 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd11a0bc-88cf-47b0-b217-e61e24c82c4a" containerName="registry-server" Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.954954 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd11a0bc-88cf-47b0-b217-e61e24c82c4a" containerName="registry-server" Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.955179 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd11a0bc-88cf-47b0-b217-e61e24c82c4a" containerName="registry-server" Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.955199 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ddf979d-a363-420d-8368-aac337f5b078" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.955964 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-khd49" Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.958626 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-c6vmh" Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.958783 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.958850 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.959131 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.959312 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 14:25:01 crc kubenswrapper[4959]: I1007 14:25:01.972995 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-khd49"] Oct 07 14:25:02 crc kubenswrapper[4959]: I1007 14:25:02.140816 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fa6bf86-611f-4c87-8475-ac0c57f5bf0a-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-khd49\" (UID: \"7fa6bf86-611f-4c87-8475-ac0c57f5bf0a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-khd49" Oct 07 14:25:02 crc kubenswrapper[4959]: I1007 14:25:02.140891 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fa6bf86-611f-4c87-8475-ac0c57f5bf0a-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-khd49\" (UID: \"7fa6bf86-611f-4c87-8475-ac0c57f5bf0a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-khd49" Oct 07 14:25:02 crc kubenswrapper[4959]: I1007 14:25:02.140938 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7fa6bf86-611f-4c87-8475-ac0c57f5bf0a-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-khd49\" (UID: \"7fa6bf86-611f-4c87-8475-ac0c57f5bf0a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-khd49" Oct 07 14:25:02 crc kubenswrapper[4959]: I1007 14:25:02.140964 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhm27\" (UniqueName: \"kubernetes.io/projected/7fa6bf86-611f-4c87-8475-ac0c57f5bf0a-kube-api-access-zhm27\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-khd49\" (UID: \"7fa6bf86-611f-4c87-8475-ac0c57f5bf0a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-khd49" Oct 07 14:25:02 crc kubenswrapper[4959]: I1007 14:25:02.242689 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fa6bf86-611f-4c87-8475-ac0c57f5bf0a-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-khd49\" (UID: \"7fa6bf86-611f-4c87-8475-ac0c57f5bf0a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-khd49" Oct 07 14:25:02 crc kubenswrapper[4959]: I1007 14:25:02.242782 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fa6bf86-611f-4c87-8475-ac0c57f5bf0a-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-khd49\" (UID: \"7fa6bf86-611f-4c87-8475-ac0c57f5bf0a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-khd49" Oct 07 14:25:02 crc kubenswrapper[4959]: I1007 14:25:02.242833 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7fa6bf86-611f-4c87-8475-ac0c57f5bf0a-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-khd49\" (UID: \"7fa6bf86-611f-4c87-8475-ac0c57f5bf0a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-khd49" Oct 07 14:25:02 crc kubenswrapper[4959]: I1007 14:25:02.242868 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhm27\" (UniqueName: \"kubernetes.io/projected/7fa6bf86-611f-4c87-8475-ac0c57f5bf0a-kube-api-access-zhm27\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-khd49\" (UID: \"7fa6bf86-611f-4c87-8475-ac0c57f5bf0a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-khd49" Oct 07 14:25:02 crc kubenswrapper[4959]: I1007 14:25:02.246959 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fa6bf86-611f-4c87-8475-ac0c57f5bf0a-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-khd49\" (UID: \"7fa6bf86-611f-4c87-8475-ac0c57f5bf0a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-khd49" Oct 07 14:25:02 crc kubenswrapper[4959]: I1007 14:25:02.247645 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fa6bf86-611f-4c87-8475-ac0c57f5bf0a-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-khd49\" (UID: \"7fa6bf86-611f-4c87-8475-ac0c57f5bf0a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-khd49" Oct 07 14:25:02 crc kubenswrapper[4959]: I1007 14:25:02.250135 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7fa6bf86-611f-4c87-8475-ac0c57f5bf0a-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-khd49\" (UID: \"7fa6bf86-611f-4c87-8475-ac0c57f5bf0a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-khd49" Oct 07 14:25:02 crc kubenswrapper[4959]: I1007 14:25:02.262131 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhm27\" (UniqueName: \"kubernetes.io/projected/7fa6bf86-611f-4c87-8475-ac0c57f5bf0a-kube-api-access-zhm27\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-khd49\" (UID: \"7fa6bf86-611f-4c87-8475-ac0c57f5bf0a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-khd49" Oct 07 14:25:02 crc kubenswrapper[4959]: I1007 14:25:02.280427 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-khd49" Oct 07 14:25:02 crc kubenswrapper[4959]: I1007 14:25:02.653996 4959 scope.go:117] "RemoveContainer" containerID="0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d" Oct 07 14:25:02 crc kubenswrapper[4959]: E1007 14:25:02.654651 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:25:02 crc kubenswrapper[4959]: I1007 14:25:02.838586 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-khd49"] Oct 07 14:25:02 crc kubenswrapper[4959]: I1007 14:25:02.865773 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-khd49" event={"ID":"7fa6bf86-611f-4c87-8475-ac0c57f5bf0a","Type":"ContainerStarted","Data":"1d6261360441d60d2d9dcf67a2a3aae0c50c7285f78c42d311712a0bdaca53e2"} Oct 07 14:25:03 crc kubenswrapper[4959]: I1007 14:25:03.876123 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-khd49" event={"ID":"7fa6bf86-611f-4c87-8475-ac0c57f5bf0a","Type":"ContainerStarted","Data":"096fafb24e31bf8c38b0b4628b22a09c37d5ef44e0c14b5244c9e1e26c54d1bc"} Oct 07 14:25:13 crc kubenswrapper[4959]: I1007 14:25:13.654097 4959 scope.go:117] "RemoveContainer" containerID="0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d" Oct 07 14:25:13 crc kubenswrapper[4959]: E1007 14:25:13.655262 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:25:24 crc kubenswrapper[4959]: I1007 14:25:24.653154 4959 scope.go:117] "RemoveContainer" containerID="0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d" Oct 07 14:25:24 crc kubenswrapper[4959]: E1007 14:25:24.654022 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:25:29 crc kubenswrapper[4959]: I1007 14:25:29.123485 4959 generic.go:334] "Generic (PLEG): container finished" podID="7fa6bf86-611f-4c87-8475-ac0c57f5bf0a" containerID="096fafb24e31bf8c38b0b4628b22a09c37d5ef44e0c14b5244c9e1e26c54d1bc" exitCode=0 Oct 07 14:25:29 crc kubenswrapper[4959]: I1007 14:25:29.123569 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-khd49" event={"ID":"7fa6bf86-611f-4c87-8475-ac0c57f5bf0a","Type":"ContainerDied","Data":"096fafb24e31bf8c38b0b4628b22a09c37d5ef44e0c14b5244c9e1e26c54d1bc"} Oct 07 14:25:30 crc kubenswrapper[4959]: I1007 14:25:30.495883 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-khd49" Oct 07 14:25:30 crc kubenswrapper[4959]: I1007 14:25:30.553667 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fa6bf86-611f-4c87-8475-ac0c57f5bf0a-ssh-key\") pod \"7fa6bf86-611f-4c87-8475-ac0c57f5bf0a\" (UID: \"7fa6bf86-611f-4c87-8475-ac0c57f5bf0a\") " Oct 07 14:25:30 crc kubenswrapper[4959]: I1007 14:25:30.553777 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fa6bf86-611f-4c87-8475-ac0c57f5bf0a-inventory\") pod \"7fa6bf86-611f-4c87-8475-ac0c57f5bf0a\" (UID: \"7fa6bf86-611f-4c87-8475-ac0c57f5bf0a\") " Oct 07 14:25:30 crc kubenswrapper[4959]: I1007 14:25:30.553824 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhm27\" (UniqueName: \"kubernetes.io/projected/7fa6bf86-611f-4c87-8475-ac0c57f5bf0a-kube-api-access-zhm27\") pod \"7fa6bf86-611f-4c87-8475-ac0c57f5bf0a\" (UID: \"7fa6bf86-611f-4c87-8475-ac0c57f5bf0a\") " Oct 07 14:25:30 crc kubenswrapper[4959]: I1007 14:25:30.553880 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7fa6bf86-611f-4c87-8475-ac0c57f5bf0a-ceph\") pod \"7fa6bf86-611f-4c87-8475-ac0c57f5bf0a\" (UID: \"7fa6bf86-611f-4c87-8475-ac0c57f5bf0a\") " Oct 07 14:25:30 crc kubenswrapper[4959]: I1007 14:25:30.560355 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fa6bf86-611f-4c87-8475-ac0c57f5bf0a-ceph" (OuterVolumeSpecName: "ceph") pod "7fa6bf86-611f-4c87-8475-ac0c57f5bf0a" (UID: "7fa6bf86-611f-4c87-8475-ac0c57f5bf0a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:25:30 crc kubenswrapper[4959]: I1007 14:25:30.583436 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fa6bf86-611f-4c87-8475-ac0c57f5bf0a-inventory" (OuterVolumeSpecName: "inventory") pod "7fa6bf86-611f-4c87-8475-ac0c57f5bf0a" (UID: "7fa6bf86-611f-4c87-8475-ac0c57f5bf0a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:25:30 crc kubenswrapper[4959]: I1007 14:25:30.584215 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fa6bf86-611f-4c87-8475-ac0c57f5bf0a-kube-api-access-zhm27" (OuterVolumeSpecName: "kube-api-access-zhm27") pod "7fa6bf86-611f-4c87-8475-ac0c57f5bf0a" (UID: "7fa6bf86-611f-4c87-8475-ac0c57f5bf0a"). InnerVolumeSpecName "kube-api-access-zhm27". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:25:30 crc kubenswrapper[4959]: I1007 14:25:30.614713 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fa6bf86-611f-4c87-8475-ac0c57f5bf0a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7fa6bf86-611f-4c87-8475-ac0c57f5bf0a" (UID: "7fa6bf86-611f-4c87-8475-ac0c57f5bf0a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:25:30 crc kubenswrapper[4959]: I1007 14:25:30.656117 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fa6bf86-611f-4c87-8475-ac0c57f5bf0a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 14:25:30 crc kubenswrapper[4959]: I1007 14:25:30.656236 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fa6bf86-611f-4c87-8475-ac0c57f5bf0a-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 14:25:30 crc kubenswrapper[4959]: I1007 14:25:30.656250 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhm27\" (UniqueName: \"kubernetes.io/projected/7fa6bf86-611f-4c87-8475-ac0c57f5bf0a-kube-api-access-zhm27\") on node \"crc\" DevicePath \"\"" Oct 07 14:25:30 crc kubenswrapper[4959]: I1007 14:25:30.656265 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7fa6bf86-611f-4c87-8475-ac0c57f5bf0a-ceph\") on node \"crc\" DevicePath \"\"" Oct 07 14:25:31 crc kubenswrapper[4959]: I1007 14:25:31.142929 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-khd49" event={"ID":"7fa6bf86-611f-4c87-8475-ac0c57f5bf0a","Type":"ContainerDied","Data":"1d6261360441d60d2d9dcf67a2a3aae0c50c7285f78c42d311712a0bdaca53e2"} Oct 07 14:25:31 crc kubenswrapper[4959]: I1007 14:25:31.142976 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d6261360441d60d2d9dcf67a2a3aae0c50c7285f78c42d311712a0bdaca53e2" Oct 07 14:25:31 crc kubenswrapper[4959]: I1007 14:25:31.143186 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-khd49" Oct 07 14:25:31 crc kubenswrapper[4959]: I1007 14:25:31.238706 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zblfq"] Oct 07 14:25:31 crc kubenswrapper[4959]: E1007 14:25:31.239167 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fa6bf86-611f-4c87-8475-ac0c57f5bf0a" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 07 14:25:31 crc kubenswrapper[4959]: I1007 14:25:31.239184 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fa6bf86-611f-4c87-8475-ac0c57f5bf0a" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 07 14:25:31 crc kubenswrapper[4959]: I1007 14:25:31.239355 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fa6bf86-611f-4c87-8475-ac0c57f5bf0a" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 07 14:25:31 crc kubenswrapper[4959]: I1007 14:25:31.240315 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zblfq" Oct 07 14:25:31 crc kubenswrapper[4959]: I1007 14:25:31.246909 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 14:25:31 crc kubenswrapper[4959]: I1007 14:25:31.247227 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 07 14:25:31 crc kubenswrapper[4959]: I1007 14:25:31.247708 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 14:25:31 crc kubenswrapper[4959]: I1007 14:25:31.247717 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 14:25:31 crc kubenswrapper[4959]: I1007 14:25:31.249444 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-c6vmh" Oct 07 14:25:31 crc kubenswrapper[4959]: I1007 14:25:31.285868 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zblfq"] Oct 07 14:25:31 crc kubenswrapper[4959]: I1007 14:25:31.374359 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f81e9c7e-9754-410a-9ac3-df9f6fcf20e9-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-zblfq\" (UID: \"f81e9c7e-9754-410a-9ac3-df9f6fcf20e9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zblfq" Oct 07 14:25:31 crc kubenswrapper[4959]: I1007 14:25:31.374473 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f81e9c7e-9754-410a-9ac3-df9f6fcf20e9-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-zblfq\" (UID: \"f81e9c7e-9754-410a-9ac3-df9f6fcf20e9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zblfq" Oct 07 14:25:31 crc kubenswrapper[4959]: I1007 14:25:31.374634 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8kbp\" (UniqueName: \"kubernetes.io/projected/f81e9c7e-9754-410a-9ac3-df9f6fcf20e9-kube-api-access-l8kbp\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-zblfq\" (UID: \"f81e9c7e-9754-410a-9ac3-df9f6fcf20e9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zblfq" Oct 07 14:25:31 crc kubenswrapper[4959]: I1007 14:25:31.375010 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f81e9c7e-9754-410a-9ac3-df9f6fcf20e9-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-zblfq\" (UID: \"f81e9c7e-9754-410a-9ac3-df9f6fcf20e9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zblfq" Oct 07 14:25:31 crc kubenswrapper[4959]: I1007 14:25:31.477922 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f81e9c7e-9754-410a-9ac3-df9f6fcf20e9-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-zblfq\" (UID: \"f81e9c7e-9754-410a-9ac3-df9f6fcf20e9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zblfq" Oct 07 14:25:31 crc kubenswrapper[4959]: I1007 14:25:31.478044 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f81e9c7e-9754-410a-9ac3-df9f6fcf20e9-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-zblfq\" (UID: \"f81e9c7e-9754-410a-9ac3-df9f6fcf20e9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zblfq" Oct 07 14:25:31 crc kubenswrapper[4959]: I1007 14:25:31.478130 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f81e9c7e-9754-410a-9ac3-df9f6fcf20e9-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-zblfq\" (UID: \"f81e9c7e-9754-410a-9ac3-df9f6fcf20e9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zblfq" Oct 07 14:25:31 crc kubenswrapper[4959]: I1007 14:25:31.478185 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8kbp\" (UniqueName: \"kubernetes.io/projected/f81e9c7e-9754-410a-9ac3-df9f6fcf20e9-kube-api-access-l8kbp\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-zblfq\" (UID: \"f81e9c7e-9754-410a-9ac3-df9f6fcf20e9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zblfq" Oct 07 14:25:31 crc kubenswrapper[4959]: I1007 14:25:31.482802 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f81e9c7e-9754-410a-9ac3-df9f6fcf20e9-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-zblfq\" (UID: \"f81e9c7e-9754-410a-9ac3-df9f6fcf20e9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zblfq" Oct 07 14:25:31 crc kubenswrapper[4959]: I1007 14:25:31.483383 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f81e9c7e-9754-410a-9ac3-df9f6fcf20e9-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-zblfq\" (UID: \"f81e9c7e-9754-410a-9ac3-df9f6fcf20e9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zblfq" Oct 07 14:25:31 crc kubenswrapper[4959]: I1007 14:25:31.485373 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f81e9c7e-9754-410a-9ac3-df9f6fcf20e9-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-zblfq\" (UID: \"f81e9c7e-9754-410a-9ac3-df9f6fcf20e9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zblfq" Oct 07 14:25:31 crc kubenswrapper[4959]: I1007 14:25:31.495396 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8kbp\" (UniqueName: \"kubernetes.io/projected/f81e9c7e-9754-410a-9ac3-df9f6fcf20e9-kube-api-access-l8kbp\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-zblfq\" (UID: \"f81e9c7e-9754-410a-9ac3-df9f6fcf20e9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zblfq" Oct 07 14:25:31 crc kubenswrapper[4959]: I1007 14:25:31.578185 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zblfq" Oct 07 14:25:32 crc kubenswrapper[4959]: I1007 14:25:32.213154 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zblfq"] Oct 07 14:25:32 crc kubenswrapper[4959]: I1007 14:25:32.214352 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 14:25:33 crc kubenswrapper[4959]: I1007 14:25:33.163329 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zblfq" event={"ID":"f81e9c7e-9754-410a-9ac3-df9f6fcf20e9","Type":"ContainerStarted","Data":"5d5f938732cb307acdf799ed716e8a219d190dc7b45489ab2ece9904687982ba"} Oct 07 14:25:34 crc kubenswrapper[4959]: I1007 14:25:34.174840 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zblfq" event={"ID":"f81e9c7e-9754-410a-9ac3-df9f6fcf20e9","Type":"ContainerStarted","Data":"ef2daf0fa9b7157932b420339d11c8d8a2b6ef56e7eceb818f929031c8fe1f80"} Oct 07 14:25:34 crc kubenswrapper[4959]: I1007 14:25:34.199981 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zblfq" podStartSLOduration=2.526470115 podStartE2EDuration="3.199953612s" podCreationTimestamp="2025-10-07 14:25:31 +0000 UTC" firstStartedPulling="2025-10-07 14:25:32.214035106 +0000 UTC m=+2414.297439431" lastFinishedPulling="2025-10-07 14:25:32.887518603 +0000 UTC m=+2414.970922928" observedRunningTime="2025-10-07 14:25:34.195002899 +0000 UTC m=+2416.278407284" watchObservedRunningTime="2025-10-07 14:25:34.199953612 +0000 UTC m=+2416.283357977" Oct 07 14:25:38 crc kubenswrapper[4959]: I1007 14:25:38.660380 4959 scope.go:117] "RemoveContainer" containerID="0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d" Oct 07 14:25:38 crc kubenswrapper[4959]: E1007 14:25:38.661443 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:25:40 crc kubenswrapper[4959]: I1007 14:25:40.230412 4959 generic.go:334] "Generic (PLEG): container finished" podID="f81e9c7e-9754-410a-9ac3-df9f6fcf20e9" containerID="ef2daf0fa9b7157932b420339d11c8d8a2b6ef56e7eceb818f929031c8fe1f80" exitCode=0 Oct 07 14:25:40 crc kubenswrapper[4959]: I1007 14:25:40.230464 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zblfq" event={"ID":"f81e9c7e-9754-410a-9ac3-df9f6fcf20e9","Type":"ContainerDied","Data":"ef2daf0fa9b7157932b420339d11c8d8a2b6ef56e7eceb818f929031c8fe1f80"} Oct 07 14:25:41 crc kubenswrapper[4959]: I1007 14:25:41.672252 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zblfq" Oct 07 14:25:41 crc kubenswrapper[4959]: I1007 14:25:41.693394 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8kbp\" (UniqueName: \"kubernetes.io/projected/f81e9c7e-9754-410a-9ac3-df9f6fcf20e9-kube-api-access-l8kbp\") pod \"f81e9c7e-9754-410a-9ac3-df9f6fcf20e9\" (UID: \"f81e9c7e-9754-410a-9ac3-df9f6fcf20e9\") " Oct 07 14:25:41 crc kubenswrapper[4959]: I1007 14:25:41.693470 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f81e9c7e-9754-410a-9ac3-df9f6fcf20e9-ceph\") pod \"f81e9c7e-9754-410a-9ac3-df9f6fcf20e9\" (UID: \"f81e9c7e-9754-410a-9ac3-df9f6fcf20e9\") " Oct 07 14:25:41 crc kubenswrapper[4959]: I1007 14:25:41.693561 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f81e9c7e-9754-410a-9ac3-df9f6fcf20e9-inventory\") pod \"f81e9c7e-9754-410a-9ac3-df9f6fcf20e9\" (UID: \"f81e9c7e-9754-410a-9ac3-df9f6fcf20e9\") " Oct 07 14:25:41 crc kubenswrapper[4959]: I1007 14:25:41.693926 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f81e9c7e-9754-410a-9ac3-df9f6fcf20e9-ssh-key\") pod \"f81e9c7e-9754-410a-9ac3-df9f6fcf20e9\" (UID: \"f81e9c7e-9754-410a-9ac3-df9f6fcf20e9\") " Oct 07 14:25:41 crc kubenswrapper[4959]: I1007 14:25:41.703308 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f81e9c7e-9754-410a-9ac3-df9f6fcf20e9-ceph" (OuterVolumeSpecName: "ceph") pod "f81e9c7e-9754-410a-9ac3-df9f6fcf20e9" (UID: "f81e9c7e-9754-410a-9ac3-df9f6fcf20e9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:25:41 crc kubenswrapper[4959]: I1007 14:25:41.703395 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f81e9c7e-9754-410a-9ac3-df9f6fcf20e9-kube-api-access-l8kbp" (OuterVolumeSpecName: "kube-api-access-l8kbp") pod "f81e9c7e-9754-410a-9ac3-df9f6fcf20e9" (UID: "f81e9c7e-9754-410a-9ac3-df9f6fcf20e9"). InnerVolumeSpecName "kube-api-access-l8kbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:25:41 crc kubenswrapper[4959]: I1007 14:25:41.721471 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f81e9c7e-9754-410a-9ac3-df9f6fcf20e9-inventory" (OuterVolumeSpecName: "inventory") pod "f81e9c7e-9754-410a-9ac3-df9f6fcf20e9" (UID: "f81e9c7e-9754-410a-9ac3-df9f6fcf20e9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:25:41 crc kubenswrapper[4959]: I1007 14:25:41.749367 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f81e9c7e-9754-410a-9ac3-df9f6fcf20e9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f81e9c7e-9754-410a-9ac3-df9f6fcf20e9" (UID: "f81e9c7e-9754-410a-9ac3-df9f6fcf20e9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:25:41 crc kubenswrapper[4959]: I1007 14:25:41.796280 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f81e9c7e-9754-410a-9ac3-df9f6fcf20e9-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 14:25:41 crc kubenswrapper[4959]: I1007 14:25:41.796321 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8kbp\" (UniqueName: \"kubernetes.io/projected/f81e9c7e-9754-410a-9ac3-df9f6fcf20e9-kube-api-access-l8kbp\") on node \"crc\" DevicePath \"\"" Oct 07 14:25:41 crc kubenswrapper[4959]: I1007 14:25:41.796335 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f81e9c7e-9754-410a-9ac3-df9f6fcf20e9-ceph\") on node \"crc\" DevicePath \"\"" Oct 07 14:25:41 crc kubenswrapper[4959]: I1007 14:25:41.796343 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f81e9c7e-9754-410a-9ac3-df9f6fcf20e9-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 14:25:42 crc kubenswrapper[4959]: I1007 14:25:42.252012 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zblfq" event={"ID":"f81e9c7e-9754-410a-9ac3-df9f6fcf20e9","Type":"ContainerDied","Data":"5d5f938732cb307acdf799ed716e8a219d190dc7b45489ab2ece9904687982ba"} Oct 07 14:25:42 crc kubenswrapper[4959]: I1007 14:25:42.252065 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d5f938732cb307acdf799ed716e8a219d190dc7b45489ab2ece9904687982ba" Oct 07 14:25:42 crc kubenswrapper[4959]: I1007 14:25:42.252093 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-zblfq" Oct 07 14:25:42 crc kubenswrapper[4959]: I1007 14:25:42.354763 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-bdlsj"] Oct 07 14:25:42 crc kubenswrapper[4959]: E1007 14:25:42.355306 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f81e9c7e-9754-410a-9ac3-df9f6fcf20e9" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 07 14:25:42 crc kubenswrapper[4959]: I1007 14:25:42.355329 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f81e9c7e-9754-410a-9ac3-df9f6fcf20e9" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 07 14:25:42 crc kubenswrapper[4959]: I1007 14:25:42.355573 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f81e9c7e-9754-410a-9ac3-df9f6fcf20e9" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 07 14:25:42 crc kubenswrapper[4959]: I1007 14:25:42.365588 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bdlsj" Oct 07 14:25:42 crc kubenswrapper[4959]: I1007 14:25:42.368575 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 07 14:25:42 crc kubenswrapper[4959]: I1007 14:25:42.370258 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-c6vmh" Oct 07 14:25:42 crc kubenswrapper[4959]: I1007 14:25:42.370284 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 14:25:42 crc kubenswrapper[4959]: I1007 14:25:42.373686 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 14:25:42 crc kubenswrapper[4959]: I1007 14:25:42.373706 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 14:25:42 crc kubenswrapper[4959]: I1007 14:25:42.386689 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-bdlsj"] Oct 07 14:25:42 crc kubenswrapper[4959]: I1007 14:25:42.409150 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99cvv\" (UniqueName: \"kubernetes.io/projected/865d8bbd-aece-4d0f-b948-a611ebe0ed6b-kube-api-access-99cvv\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bdlsj\" (UID: \"865d8bbd-aece-4d0f-b948-a611ebe0ed6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bdlsj" Oct 07 14:25:42 crc kubenswrapper[4959]: I1007 14:25:42.409207 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/865d8bbd-aece-4d0f-b948-a611ebe0ed6b-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bdlsj\" (UID: \"865d8bbd-aece-4d0f-b948-a611ebe0ed6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bdlsj" Oct 07 14:25:42 crc kubenswrapper[4959]: I1007 14:25:42.409366 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/865d8bbd-aece-4d0f-b948-a611ebe0ed6b-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bdlsj\" (UID: \"865d8bbd-aece-4d0f-b948-a611ebe0ed6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bdlsj" Oct 07 14:25:42 crc kubenswrapper[4959]: I1007 14:25:42.409408 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/865d8bbd-aece-4d0f-b948-a611ebe0ed6b-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bdlsj\" (UID: \"865d8bbd-aece-4d0f-b948-a611ebe0ed6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bdlsj" Oct 07 14:25:42 crc kubenswrapper[4959]: I1007 14:25:42.511319 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/865d8bbd-aece-4d0f-b948-a611ebe0ed6b-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bdlsj\" (UID: \"865d8bbd-aece-4d0f-b948-a611ebe0ed6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bdlsj" Oct 07 14:25:42 crc kubenswrapper[4959]: I1007 14:25:42.511643 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/865d8bbd-aece-4d0f-b948-a611ebe0ed6b-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bdlsj\" (UID: \"865d8bbd-aece-4d0f-b948-a611ebe0ed6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bdlsj" Oct 07 14:25:42 crc kubenswrapper[4959]: I1007 14:25:42.511705 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99cvv\" (UniqueName: \"kubernetes.io/projected/865d8bbd-aece-4d0f-b948-a611ebe0ed6b-kube-api-access-99cvv\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bdlsj\" (UID: \"865d8bbd-aece-4d0f-b948-a611ebe0ed6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bdlsj" Oct 07 14:25:42 crc kubenswrapper[4959]: I1007 14:25:42.511733 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/865d8bbd-aece-4d0f-b948-a611ebe0ed6b-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bdlsj\" (UID: \"865d8bbd-aece-4d0f-b948-a611ebe0ed6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bdlsj" Oct 07 14:25:42 crc kubenswrapper[4959]: I1007 14:25:42.515636 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/865d8bbd-aece-4d0f-b948-a611ebe0ed6b-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bdlsj\" (UID: \"865d8bbd-aece-4d0f-b948-a611ebe0ed6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bdlsj" Oct 07 14:25:42 crc kubenswrapper[4959]: I1007 14:25:42.516074 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/865d8bbd-aece-4d0f-b948-a611ebe0ed6b-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bdlsj\" (UID: \"865d8bbd-aece-4d0f-b948-a611ebe0ed6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bdlsj" Oct 07 14:25:42 crc kubenswrapper[4959]: I1007 14:25:42.516494 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/865d8bbd-aece-4d0f-b948-a611ebe0ed6b-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bdlsj\" (UID: \"865d8bbd-aece-4d0f-b948-a611ebe0ed6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bdlsj" Oct 07 14:25:42 crc kubenswrapper[4959]: I1007 14:25:42.527229 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99cvv\" (UniqueName: \"kubernetes.io/projected/865d8bbd-aece-4d0f-b948-a611ebe0ed6b-kube-api-access-99cvv\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bdlsj\" (UID: \"865d8bbd-aece-4d0f-b948-a611ebe0ed6b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bdlsj" Oct 07 14:25:42 crc kubenswrapper[4959]: I1007 14:25:42.685397 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bdlsj" Oct 07 14:25:43 crc kubenswrapper[4959]: I1007 14:25:43.182036 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-bdlsj"] Oct 07 14:25:43 crc kubenswrapper[4959]: I1007 14:25:43.260580 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bdlsj" event={"ID":"865d8bbd-aece-4d0f-b948-a611ebe0ed6b","Type":"ContainerStarted","Data":"7e010cd5ceccda0af90cc74e27e94fd862bc279cd1bb5334308f01d74211891a"} Oct 07 14:25:44 crc kubenswrapper[4959]: I1007 14:25:44.270621 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bdlsj" event={"ID":"865d8bbd-aece-4d0f-b948-a611ebe0ed6b","Type":"ContainerStarted","Data":"93052fb56b0410bfe296ff8660ed4f4bb15228320a9feb84bf5f0baf6d124c33"} Oct 07 14:25:44 crc kubenswrapper[4959]: I1007 14:25:44.290031 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bdlsj" podStartSLOduration=1.870652813 podStartE2EDuration="2.290013147s" podCreationTimestamp="2025-10-07 14:25:42 +0000 UTC" firstStartedPulling="2025-10-07 14:25:43.185123598 +0000 UTC m=+2425.268527923" lastFinishedPulling="2025-10-07 14:25:43.604483932 +0000 UTC m=+2425.687888257" observedRunningTime="2025-10-07 14:25:44.285771402 +0000 UTC m=+2426.369175727" watchObservedRunningTime="2025-10-07 14:25:44.290013147 +0000 UTC m=+2426.373417472" Oct 07 14:25:50 crc kubenswrapper[4959]: I1007 14:25:50.654890 4959 scope.go:117] "RemoveContainer" containerID="0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d" Oct 07 14:25:50 crc kubenswrapper[4959]: E1007 14:25:50.655511 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:26:04 crc kubenswrapper[4959]: I1007 14:26:04.654921 4959 scope.go:117] "RemoveContainer" containerID="0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d" Oct 07 14:26:04 crc kubenswrapper[4959]: E1007 14:26:04.656383 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:26:15 crc kubenswrapper[4959]: I1007 14:26:15.653484 4959 scope.go:117] "RemoveContainer" containerID="0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d" Oct 07 14:26:15 crc kubenswrapper[4959]: E1007 14:26:15.654421 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:26:21 crc kubenswrapper[4959]: I1007 14:26:21.597689 4959 generic.go:334] "Generic (PLEG): container finished" podID="865d8bbd-aece-4d0f-b948-a611ebe0ed6b" containerID="93052fb56b0410bfe296ff8660ed4f4bb15228320a9feb84bf5f0baf6d124c33" exitCode=0 Oct 07 14:26:21 crc kubenswrapper[4959]: I1007 14:26:21.597805 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bdlsj" event={"ID":"865d8bbd-aece-4d0f-b948-a611ebe0ed6b","Type":"ContainerDied","Data":"93052fb56b0410bfe296ff8660ed4f4bb15228320a9feb84bf5f0baf6d124c33"} Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.058758 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bdlsj" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.104268 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/865d8bbd-aece-4d0f-b948-a611ebe0ed6b-ceph\") pod \"865d8bbd-aece-4d0f-b948-a611ebe0ed6b\" (UID: \"865d8bbd-aece-4d0f-b948-a611ebe0ed6b\") " Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.104346 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/865d8bbd-aece-4d0f-b948-a611ebe0ed6b-ssh-key\") pod \"865d8bbd-aece-4d0f-b948-a611ebe0ed6b\" (UID: \"865d8bbd-aece-4d0f-b948-a611ebe0ed6b\") " Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.104485 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/865d8bbd-aece-4d0f-b948-a611ebe0ed6b-inventory\") pod \"865d8bbd-aece-4d0f-b948-a611ebe0ed6b\" (UID: \"865d8bbd-aece-4d0f-b948-a611ebe0ed6b\") " Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.104545 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99cvv\" (UniqueName: \"kubernetes.io/projected/865d8bbd-aece-4d0f-b948-a611ebe0ed6b-kube-api-access-99cvv\") pod \"865d8bbd-aece-4d0f-b948-a611ebe0ed6b\" (UID: \"865d8bbd-aece-4d0f-b948-a611ebe0ed6b\") " Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.113409 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/865d8bbd-aece-4d0f-b948-a611ebe0ed6b-ceph" (OuterVolumeSpecName: "ceph") pod "865d8bbd-aece-4d0f-b948-a611ebe0ed6b" (UID: "865d8bbd-aece-4d0f-b948-a611ebe0ed6b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.114034 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/865d8bbd-aece-4d0f-b948-a611ebe0ed6b-kube-api-access-99cvv" (OuterVolumeSpecName: "kube-api-access-99cvv") pod "865d8bbd-aece-4d0f-b948-a611ebe0ed6b" (UID: "865d8bbd-aece-4d0f-b948-a611ebe0ed6b"). InnerVolumeSpecName "kube-api-access-99cvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.139341 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/865d8bbd-aece-4d0f-b948-a611ebe0ed6b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "865d8bbd-aece-4d0f-b948-a611ebe0ed6b" (UID: "865d8bbd-aece-4d0f-b948-a611ebe0ed6b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.143944 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/865d8bbd-aece-4d0f-b948-a611ebe0ed6b-inventory" (OuterVolumeSpecName: "inventory") pod "865d8bbd-aece-4d0f-b948-a611ebe0ed6b" (UID: "865d8bbd-aece-4d0f-b948-a611ebe0ed6b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.207963 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/865d8bbd-aece-4d0f-b948-a611ebe0ed6b-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.208020 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99cvv\" (UniqueName: \"kubernetes.io/projected/865d8bbd-aece-4d0f-b948-a611ebe0ed6b-kube-api-access-99cvv\") on node \"crc\" DevicePath \"\"" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.208037 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/865d8bbd-aece-4d0f-b948-a611ebe0ed6b-ceph\") on node \"crc\" DevicePath \"\"" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.208052 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/865d8bbd-aece-4d0f-b948-a611ebe0ed6b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.634026 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bdlsj" event={"ID":"865d8bbd-aece-4d0f-b948-a611ebe0ed6b","Type":"ContainerDied","Data":"7e010cd5ceccda0af90cc74e27e94fd862bc279cd1bb5334308f01d74211891a"} Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.634167 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bdlsj" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.634200 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e010cd5ceccda0af90cc74e27e94fd862bc279cd1bb5334308f01d74211891a" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.747730 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj"] Oct 07 14:26:23 crc kubenswrapper[4959]: E1007 14:26:23.748330 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="865d8bbd-aece-4d0f-b948-a611ebe0ed6b" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.748358 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="865d8bbd-aece-4d0f-b948-a611ebe0ed6b" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.748650 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="865d8bbd-aece-4d0f-b948-a611ebe0ed6b" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.749596 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.752195 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.752338 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-c6vmh" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.753266 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.755207 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.755419 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.762469 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj"] Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.819703 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jw8jj\" (UniqueName: \"kubernetes.io/projected/2325565a-d2ae-45a1-bece-46cda83edc36-kube-api-access-jw8jj\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj\" (UID: \"2325565a-d2ae-45a1-bece-46cda83edc36\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.819843 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2325565a-d2ae-45a1-bece-46cda83edc36-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj\" (UID: \"2325565a-d2ae-45a1-bece-46cda83edc36\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.819889 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2325565a-d2ae-45a1-bece-46cda83edc36-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj\" (UID: \"2325565a-d2ae-45a1-bece-46cda83edc36\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.819912 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2325565a-d2ae-45a1-bece-46cda83edc36-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj\" (UID: \"2325565a-d2ae-45a1-bece-46cda83edc36\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.921307 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2325565a-d2ae-45a1-bece-46cda83edc36-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj\" (UID: \"2325565a-d2ae-45a1-bece-46cda83edc36\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.921365 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2325565a-d2ae-45a1-bece-46cda83edc36-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj\" (UID: \"2325565a-d2ae-45a1-bece-46cda83edc36\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.921486 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jw8jj\" (UniqueName: \"kubernetes.io/projected/2325565a-d2ae-45a1-bece-46cda83edc36-kube-api-access-jw8jj\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj\" (UID: \"2325565a-d2ae-45a1-bece-46cda83edc36\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.921566 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2325565a-d2ae-45a1-bece-46cda83edc36-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj\" (UID: \"2325565a-d2ae-45a1-bece-46cda83edc36\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.927592 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2325565a-d2ae-45a1-bece-46cda83edc36-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj\" (UID: \"2325565a-d2ae-45a1-bece-46cda83edc36\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.930049 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2325565a-d2ae-45a1-bece-46cda83edc36-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj\" (UID: \"2325565a-d2ae-45a1-bece-46cda83edc36\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.935823 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2325565a-d2ae-45a1-bece-46cda83edc36-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj\" (UID: \"2325565a-d2ae-45a1-bece-46cda83edc36\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj" Oct 07 14:26:23 crc kubenswrapper[4959]: I1007 14:26:23.944501 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jw8jj\" (UniqueName: \"kubernetes.io/projected/2325565a-d2ae-45a1-bece-46cda83edc36-kube-api-access-jw8jj\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj\" (UID: \"2325565a-d2ae-45a1-bece-46cda83edc36\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj" Oct 07 14:26:24 crc kubenswrapper[4959]: I1007 14:26:24.082754 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj" Oct 07 14:26:24 crc kubenswrapper[4959]: I1007 14:26:24.733803 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj"] Oct 07 14:26:25 crc kubenswrapper[4959]: I1007 14:26:25.664489 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj" event={"ID":"2325565a-d2ae-45a1-bece-46cda83edc36","Type":"ContainerStarted","Data":"ddf88278a70581d52937b20068fc5da5770e65b2467c41de07d0c4f2a0385f57"} Oct 07 14:26:25 crc kubenswrapper[4959]: I1007 14:26:25.665422 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj" event={"ID":"2325565a-d2ae-45a1-bece-46cda83edc36","Type":"ContainerStarted","Data":"8ff2240c52ed48064499bb4dbfee9f7a110264ac3f9a7f2528d17d6c12f10e93"} Oct 07 14:26:25 crc kubenswrapper[4959]: I1007 14:26:25.693432 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj" podStartSLOduration=2.219107873 podStartE2EDuration="2.693406027s" podCreationTimestamp="2025-10-07 14:26:23 +0000 UTC" firstStartedPulling="2025-10-07 14:26:24.733838972 +0000 UTC m=+2466.817243307" lastFinishedPulling="2025-10-07 14:26:25.208137136 +0000 UTC m=+2467.291541461" observedRunningTime="2025-10-07 14:26:25.686630536 +0000 UTC m=+2467.770034891" watchObservedRunningTime="2025-10-07 14:26:25.693406027 +0000 UTC m=+2467.776810352" Oct 07 14:26:26 crc kubenswrapper[4959]: I1007 14:26:26.654050 4959 scope.go:117] "RemoveContainer" containerID="0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d" Oct 07 14:26:26 crc kubenswrapper[4959]: E1007 14:26:26.654378 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:26:30 crc kubenswrapper[4959]: I1007 14:26:30.737612 4959 generic.go:334] "Generic (PLEG): container finished" podID="2325565a-d2ae-45a1-bece-46cda83edc36" containerID="ddf88278a70581d52937b20068fc5da5770e65b2467c41de07d0c4f2a0385f57" exitCode=0 Oct 07 14:26:30 crc kubenswrapper[4959]: I1007 14:26:30.738290 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj" event={"ID":"2325565a-d2ae-45a1-bece-46cda83edc36","Type":"ContainerDied","Data":"ddf88278a70581d52937b20068fc5da5770e65b2467c41de07d0c4f2a0385f57"} Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.223297 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj" Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.336959 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2325565a-d2ae-45a1-bece-46cda83edc36-ceph\") pod \"2325565a-d2ae-45a1-bece-46cda83edc36\" (UID: \"2325565a-d2ae-45a1-bece-46cda83edc36\") " Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.337031 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2325565a-d2ae-45a1-bece-46cda83edc36-ssh-key\") pod \"2325565a-d2ae-45a1-bece-46cda83edc36\" (UID: \"2325565a-d2ae-45a1-bece-46cda83edc36\") " Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.337231 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jw8jj\" (UniqueName: \"kubernetes.io/projected/2325565a-d2ae-45a1-bece-46cda83edc36-kube-api-access-jw8jj\") pod \"2325565a-d2ae-45a1-bece-46cda83edc36\" (UID: \"2325565a-d2ae-45a1-bece-46cda83edc36\") " Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.337280 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2325565a-d2ae-45a1-bece-46cda83edc36-inventory\") pod \"2325565a-d2ae-45a1-bece-46cda83edc36\" (UID: \"2325565a-d2ae-45a1-bece-46cda83edc36\") " Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.344531 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2325565a-d2ae-45a1-bece-46cda83edc36-ceph" (OuterVolumeSpecName: "ceph") pod "2325565a-d2ae-45a1-bece-46cda83edc36" (UID: "2325565a-d2ae-45a1-bece-46cda83edc36"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.358409 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2325565a-d2ae-45a1-bece-46cda83edc36-kube-api-access-jw8jj" (OuterVolumeSpecName: "kube-api-access-jw8jj") pod "2325565a-d2ae-45a1-bece-46cda83edc36" (UID: "2325565a-d2ae-45a1-bece-46cda83edc36"). InnerVolumeSpecName "kube-api-access-jw8jj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.383442 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2325565a-d2ae-45a1-bece-46cda83edc36-inventory" (OuterVolumeSpecName: "inventory") pod "2325565a-d2ae-45a1-bece-46cda83edc36" (UID: "2325565a-d2ae-45a1-bece-46cda83edc36"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.400295 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2325565a-d2ae-45a1-bece-46cda83edc36-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2325565a-d2ae-45a1-bece-46cda83edc36" (UID: "2325565a-d2ae-45a1-bece-46cda83edc36"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.440815 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jw8jj\" (UniqueName: \"kubernetes.io/projected/2325565a-d2ae-45a1-bece-46cda83edc36-kube-api-access-jw8jj\") on node \"crc\" DevicePath \"\"" Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.440863 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2325565a-d2ae-45a1-bece-46cda83edc36-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.440876 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2325565a-d2ae-45a1-bece-46cda83edc36-ceph\") on node \"crc\" DevicePath \"\"" Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.440890 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2325565a-d2ae-45a1-bece-46cda83edc36-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.766975 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj" event={"ID":"2325565a-d2ae-45a1-bece-46cda83edc36","Type":"ContainerDied","Data":"8ff2240c52ed48064499bb4dbfee9f7a110264ac3f9a7f2528d17d6c12f10e93"} Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.767025 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj" Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.767057 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ff2240c52ed48064499bb4dbfee9f7a110264ac3f9a7f2528d17d6c12f10e93" Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.865612 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv"] Oct 07 14:26:32 crc kubenswrapper[4959]: E1007 14:26:32.866212 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2325565a-d2ae-45a1-bece-46cda83edc36" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.866284 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2325565a-d2ae-45a1-bece-46cda83edc36" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.866584 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2325565a-d2ae-45a1-bece-46cda83edc36" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.867567 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv" Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.872992 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.873408 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.876744 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.879092 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-c6vmh" Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.879312 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.888895 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv"] Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.949587 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/57e3b6ad-f589-41cc-8ba5-8009fe0ab434-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv\" (UID: \"57e3b6ad-f589-41cc-8ba5-8009fe0ab434\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv" Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.949645 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6xlc\" (UniqueName: \"kubernetes.io/projected/57e3b6ad-f589-41cc-8ba5-8009fe0ab434-kube-api-access-p6xlc\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv\" (UID: \"57e3b6ad-f589-41cc-8ba5-8009fe0ab434\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv" Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.949934 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57e3b6ad-f589-41cc-8ba5-8009fe0ab434-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv\" (UID: \"57e3b6ad-f589-41cc-8ba5-8009fe0ab434\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv" Oct 07 14:26:32 crc kubenswrapper[4959]: I1007 14:26:32.950315 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57e3b6ad-f589-41cc-8ba5-8009fe0ab434-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv\" (UID: \"57e3b6ad-f589-41cc-8ba5-8009fe0ab434\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv" Oct 07 14:26:33 crc kubenswrapper[4959]: I1007 14:26:33.052687 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/57e3b6ad-f589-41cc-8ba5-8009fe0ab434-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv\" (UID: \"57e3b6ad-f589-41cc-8ba5-8009fe0ab434\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv" Oct 07 14:26:33 crc kubenswrapper[4959]: I1007 14:26:33.052746 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6xlc\" (UniqueName: \"kubernetes.io/projected/57e3b6ad-f589-41cc-8ba5-8009fe0ab434-kube-api-access-p6xlc\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv\" (UID: \"57e3b6ad-f589-41cc-8ba5-8009fe0ab434\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv" Oct 07 14:26:33 crc kubenswrapper[4959]: I1007 14:26:33.052810 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57e3b6ad-f589-41cc-8ba5-8009fe0ab434-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv\" (UID: \"57e3b6ad-f589-41cc-8ba5-8009fe0ab434\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv" Oct 07 14:26:33 crc kubenswrapper[4959]: I1007 14:26:33.052907 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57e3b6ad-f589-41cc-8ba5-8009fe0ab434-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv\" (UID: \"57e3b6ad-f589-41cc-8ba5-8009fe0ab434\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv" Oct 07 14:26:33 crc kubenswrapper[4959]: I1007 14:26:33.058148 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/57e3b6ad-f589-41cc-8ba5-8009fe0ab434-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv\" (UID: \"57e3b6ad-f589-41cc-8ba5-8009fe0ab434\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv" Oct 07 14:26:33 crc kubenswrapper[4959]: I1007 14:26:33.058857 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57e3b6ad-f589-41cc-8ba5-8009fe0ab434-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv\" (UID: \"57e3b6ad-f589-41cc-8ba5-8009fe0ab434\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv" Oct 07 14:26:33 crc kubenswrapper[4959]: I1007 14:26:33.060170 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57e3b6ad-f589-41cc-8ba5-8009fe0ab434-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv\" (UID: \"57e3b6ad-f589-41cc-8ba5-8009fe0ab434\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv" Oct 07 14:26:33 crc kubenswrapper[4959]: I1007 14:26:33.085769 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6xlc\" (UniqueName: \"kubernetes.io/projected/57e3b6ad-f589-41cc-8ba5-8009fe0ab434-kube-api-access-p6xlc\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv\" (UID: \"57e3b6ad-f589-41cc-8ba5-8009fe0ab434\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv" Oct 07 14:26:33 crc kubenswrapper[4959]: I1007 14:26:33.227359 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv" Oct 07 14:26:33 crc kubenswrapper[4959]: I1007 14:26:33.607882 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv"] Oct 07 14:26:33 crc kubenswrapper[4959]: I1007 14:26:33.776566 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv" event={"ID":"57e3b6ad-f589-41cc-8ba5-8009fe0ab434","Type":"ContainerStarted","Data":"61427feb47c42328bb552f60bf971c68f59e22c54c2b63fbb5d830049c402ad0"} Oct 07 14:26:35 crc kubenswrapper[4959]: I1007 14:26:35.803472 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv" event={"ID":"57e3b6ad-f589-41cc-8ba5-8009fe0ab434","Type":"ContainerStarted","Data":"e1935386045cbd9aa7d1cdf124c37897b7d991e14dece809e59198173ce9a2fc"} Oct 07 14:26:35 crc kubenswrapper[4959]: I1007 14:26:35.836889 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv" podStartSLOduration=2.798611609 podStartE2EDuration="3.836861835s" podCreationTimestamp="2025-10-07 14:26:32 +0000 UTC" firstStartedPulling="2025-10-07 14:26:33.623690273 +0000 UTC m=+2475.707094598" lastFinishedPulling="2025-10-07 14:26:34.661940499 +0000 UTC m=+2476.745344824" observedRunningTime="2025-10-07 14:26:35.829525229 +0000 UTC m=+2477.912929564" watchObservedRunningTime="2025-10-07 14:26:35.836861835 +0000 UTC m=+2477.920266170" Oct 07 14:26:38 crc kubenswrapper[4959]: I1007 14:26:38.665553 4959 scope.go:117] "RemoveContainer" containerID="0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d" Oct 07 14:26:38 crc kubenswrapper[4959]: E1007 14:26:38.666752 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:26:51 crc kubenswrapper[4959]: I1007 14:26:51.655480 4959 scope.go:117] "RemoveContainer" containerID="0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d" Oct 07 14:26:51 crc kubenswrapper[4959]: E1007 14:26:51.656493 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:27:02 crc kubenswrapper[4959]: I1007 14:27:02.654581 4959 scope.go:117] "RemoveContainer" containerID="0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d" Oct 07 14:27:02 crc kubenswrapper[4959]: E1007 14:27:02.656394 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:27:14 crc kubenswrapper[4959]: I1007 14:27:14.654745 4959 scope.go:117] "RemoveContainer" containerID="0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d" Oct 07 14:27:14 crc kubenswrapper[4959]: E1007 14:27:14.656137 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:27:21 crc kubenswrapper[4959]: I1007 14:27:21.263040 4959 generic.go:334] "Generic (PLEG): container finished" podID="57e3b6ad-f589-41cc-8ba5-8009fe0ab434" containerID="e1935386045cbd9aa7d1cdf124c37897b7d991e14dece809e59198173ce9a2fc" exitCode=0 Oct 07 14:27:21 crc kubenswrapper[4959]: I1007 14:27:21.263196 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv" event={"ID":"57e3b6ad-f589-41cc-8ba5-8009fe0ab434","Type":"ContainerDied","Data":"e1935386045cbd9aa7d1cdf124c37897b7d991e14dece809e59198173ce9a2fc"} Oct 07 14:27:22 crc kubenswrapper[4959]: I1007 14:27:22.717984 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv" Oct 07 14:27:22 crc kubenswrapper[4959]: I1007 14:27:22.832991 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/57e3b6ad-f589-41cc-8ba5-8009fe0ab434-ceph\") pod \"57e3b6ad-f589-41cc-8ba5-8009fe0ab434\" (UID: \"57e3b6ad-f589-41cc-8ba5-8009fe0ab434\") " Oct 07 14:27:22 crc kubenswrapper[4959]: I1007 14:27:22.833152 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6xlc\" (UniqueName: \"kubernetes.io/projected/57e3b6ad-f589-41cc-8ba5-8009fe0ab434-kube-api-access-p6xlc\") pod \"57e3b6ad-f589-41cc-8ba5-8009fe0ab434\" (UID: \"57e3b6ad-f589-41cc-8ba5-8009fe0ab434\") " Oct 07 14:27:22 crc kubenswrapper[4959]: I1007 14:27:22.833196 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57e3b6ad-f589-41cc-8ba5-8009fe0ab434-ssh-key\") pod \"57e3b6ad-f589-41cc-8ba5-8009fe0ab434\" (UID: \"57e3b6ad-f589-41cc-8ba5-8009fe0ab434\") " Oct 07 14:27:22 crc kubenswrapper[4959]: I1007 14:27:22.833303 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57e3b6ad-f589-41cc-8ba5-8009fe0ab434-inventory\") pod \"57e3b6ad-f589-41cc-8ba5-8009fe0ab434\" (UID: \"57e3b6ad-f589-41cc-8ba5-8009fe0ab434\") " Oct 07 14:27:22 crc kubenswrapper[4959]: I1007 14:27:22.839602 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57e3b6ad-f589-41cc-8ba5-8009fe0ab434-ceph" (OuterVolumeSpecName: "ceph") pod "57e3b6ad-f589-41cc-8ba5-8009fe0ab434" (UID: "57e3b6ad-f589-41cc-8ba5-8009fe0ab434"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:27:22 crc kubenswrapper[4959]: I1007 14:27:22.840662 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57e3b6ad-f589-41cc-8ba5-8009fe0ab434-kube-api-access-p6xlc" (OuterVolumeSpecName: "kube-api-access-p6xlc") pod "57e3b6ad-f589-41cc-8ba5-8009fe0ab434" (UID: "57e3b6ad-f589-41cc-8ba5-8009fe0ab434"). InnerVolumeSpecName "kube-api-access-p6xlc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:27:22 crc kubenswrapper[4959]: I1007 14:27:22.864500 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57e3b6ad-f589-41cc-8ba5-8009fe0ab434-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "57e3b6ad-f589-41cc-8ba5-8009fe0ab434" (UID: "57e3b6ad-f589-41cc-8ba5-8009fe0ab434"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:27:22 crc kubenswrapper[4959]: I1007 14:27:22.864568 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57e3b6ad-f589-41cc-8ba5-8009fe0ab434-inventory" (OuterVolumeSpecName: "inventory") pod "57e3b6ad-f589-41cc-8ba5-8009fe0ab434" (UID: "57e3b6ad-f589-41cc-8ba5-8009fe0ab434"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:27:22 crc kubenswrapper[4959]: I1007 14:27:22.935958 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/57e3b6ad-f589-41cc-8ba5-8009fe0ab434-ceph\") on node \"crc\" DevicePath \"\"" Oct 07 14:27:22 crc kubenswrapper[4959]: I1007 14:27:22.935996 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6xlc\" (UniqueName: \"kubernetes.io/projected/57e3b6ad-f589-41cc-8ba5-8009fe0ab434-kube-api-access-p6xlc\") on node \"crc\" DevicePath \"\"" Oct 07 14:27:22 crc kubenswrapper[4959]: I1007 14:27:22.936006 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57e3b6ad-f589-41cc-8ba5-8009fe0ab434-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 14:27:22 crc kubenswrapper[4959]: I1007 14:27:22.936017 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57e3b6ad-f589-41cc-8ba5-8009fe0ab434-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 14:27:23 crc kubenswrapper[4959]: I1007 14:27:23.288926 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv" event={"ID":"57e3b6ad-f589-41cc-8ba5-8009fe0ab434","Type":"ContainerDied","Data":"61427feb47c42328bb552f60bf971c68f59e22c54c2b63fbb5d830049c402ad0"} Oct 07 14:27:23 crc kubenswrapper[4959]: I1007 14:27:23.288980 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61427feb47c42328bb552f60bf971c68f59e22c54c2b63fbb5d830049c402ad0" Oct 07 14:27:23 crc kubenswrapper[4959]: I1007 14:27:23.289044 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv" Oct 07 14:27:23 crc kubenswrapper[4959]: I1007 14:27:23.398231 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ms8vj"] Oct 07 14:27:23 crc kubenswrapper[4959]: E1007 14:27:23.398742 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57e3b6ad-f589-41cc-8ba5-8009fe0ab434" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 07 14:27:23 crc kubenswrapper[4959]: I1007 14:27:23.398769 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="57e3b6ad-f589-41cc-8ba5-8009fe0ab434" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 07 14:27:23 crc kubenswrapper[4959]: I1007 14:27:23.398951 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="57e3b6ad-f589-41cc-8ba5-8009fe0ab434" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 07 14:27:23 crc kubenswrapper[4959]: I1007 14:27:23.399575 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ms8vj" Oct 07 14:27:23 crc kubenswrapper[4959]: I1007 14:27:23.403785 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 14:27:23 crc kubenswrapper[4959]: I1007 14:27:23.404011 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 14:27:23 crc kubenswrapper[4959]: I1007 14:27:23.404277 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 14:27:23 crc kubenswrapper[4959]: I1007 14:27:23.404545 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-c6vmh" Oct 07 14:27:23 crc kubenswrapper[4959]: I1007 14:27:23.405238 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 07 14:27:23 crc kubenswrapper[4959]: I1007 14:27:23.408359 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ms8vj"] Oct 07 14:27:23 crc kubenswrapper[4959]: I1007 14:27:23.549731 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/634f1665-e210-4c71-9c7a-8cf30b5f3fec-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ms8vj\" (UID: \"634f1665-e210-4c71-9c7a-8cf30b5f3fec\") " pod="openstack/ssh-known-hosts-edpm-deployment-ms8vj" Oct 07 14:27:23 crc kubenswrapper[4959]: I1007 14:27:23.549963 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/634f1665-e210-4c71-9c7a-8cf30b5f3fec-ceph\") pod \"ssh-known-hosts-edpm-deployment-ms8vj\" (UID: \"634f1665-e210-4c71-9c7a-8cf30b5f3fec\") " pod="openstack/ssh-known-hosts-edpm-deployment-ms8vj" Oct 07 14:27:23 crc kubenswrapper[4959]: I1007 14:27:23.550579 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/634f1665-e210-4c71-9c7a-8cf30b5f3fec-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ms8vj\" (UID: \"634f1665-e210-4c71-9c7a-8cf30b5f3fec\") " pod="openstack/ssh-known-hosts-edpm-deployment-ms8vj" Oct 07 14:27:23 crc kubenswrapper[4959]: I1007 14:27:23.550741 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6hhg\" (UniqueName: \"kubernetes.io/projected/634f1665-e210-4c71-9c7a-8cf30b5f3fec-kube-api-access-g6hhg\") pod \"ssh-known-hosts-edpm-deployment-ms8vj\" (UID: \"634f1665-e210-4c71-9c7a-8cf30b5f3fec\") " pod="openstack/ssh-known-hosts-edpm-deployment-ms8vj" Oct 07 14:27:23 crc kubenswrapper[4959]: I1007 14:27:23.653172 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/634f1665-e210-4c71-9c7a-8cf30b5f3fec-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ms8vj\" (UID: \"634f1665-e210-4c71-9c7a-8cf30b5f3fec\") " pod="openstack/ssh-known-hosts-edpm-deployment-ms8vj" Oct 07 14:27:23 crc kubenswrapper[4959]: I1007 14:27:23.653253 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/634f1665-e210-4c71-9c7a-8cf30b5f3fec-ceph\") pod \"ssh-known-hosts-edpm-deployment-ms8vj\" (UID: \"634f1665-e210-4c71-9c7a-8cf30b5f3fec\") " pod="openstack/ssh-known-hosts-edpm-deployment-ms8vj" Oct 07 14:27:23 crc kubenswrapper[4959]: I1007 14:27:23.653329 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/634f1665-e210-4c71-9c7a-8cf30b5f3fec-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ms8vj\" (UID: \"634f1665-e210-4c71-9c7a-8cf30b5f3fec\") " pod="openstack/ssh-known-hosts-edpm-deployment-ms8vj" Oct 07 14:27:23 crc kubenswrapper[4959]: I1007 14:27:23.653360 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6hhg\" (UniqueName: \"kubernetes.io/projected/634f1665-e210-4c71-9c7a-8cf30b5f3fec-kube-api-access-g6hhg\") pod \"ssh-known-hosts-edpm-deployment-ms8vj\" (UID: \"634f1665-e210-4c71-9c7a-8cf30b5f3fec\") " pod="openstack/ssh-known-hosts-edpm-deployment-ms8vj" Oct 07 14:27:23 crc kubenswrapper[4959]: I1007 14:27:23.660535 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/634f1665-e210-4c71-9c7a-8cf30b5f3fec-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ms8vj\" (UID: \"634f1665-e210-4c71-9c7a-8cf30b5f3fec\") " pod="openstack/ssh-known-hosts-edpm-deployment-ms8vj" Oct 07 14:27:23 crc kubenswrapper[4959]: I1007 14:27:23.660708 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/634f1665-e210-4c71-9c7a-8cf30b5f3fec-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ms8vj\" (UID: \"634f1665-e210-4c71-9c7a-8cf30b5f3fec\") " pod="openstack/ssh-known-hosts-edpm-deployment-ms8vj" Oct 07 14:27:23 crc kubenswrapper[4959]: I1007 14:27:23.660972 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/634f1665-e210-4c71-9c7a-8cf30b5f3fec-ceph\") pod \"ssh-known-hosts-edpm-deployment-ms8vj\" (UID: \"634f1665-e210-4c71-9c7a-8cf30b5f3fec\") " pod="openstack/ssh-known-hosts-edpm-deployment-ms8vj" Oct 07 14:27:23 crc kubenswrapper[4959]: I1007 14:27:23.679556 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6hhg\" (UniqueName: \"kubernetes.io/projected/634f1665-e210-4c71-9c7a-8cf30b5f3fec-kube-api-access-g6hhg\") pod \"ssh-known-hosts-edpm-deployment-ms8vj\" (UID: \"634f1665-e210-4c71-9c7a-8cf30b5f3fec\") " pod="openstack/ssh-known-hosts-edpm-deployment-ms8vj" Oct 07 14:27:23 crc kubenswrapper[4959]: I1007 14:27:23.724888 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ms8vj" Oct 07 14:27:24 crc kubenswrapper[4959]: I1007 14:27:24.308793 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ms8vj"] Oct 07 14:27:25 crc kubenswrapper[4959]: I1007 14:27:25.346240 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ms8vj" event={"ID":"634f1665-e210-4c71-9c7a-8cf30b5f3fec","Type":"ContainerStarted","Data":"92b8c54c7449e7c3bfe20512c60b67fff9f168b70ae22f0171766d931d29b264"} Oct 07 14:27:25 crc kubenswrapper[4959]: I1007 14:27:25.346681 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ms8vj" event={"ID":"634f1665-e210-4c71-9c7a-8cf30b5f3fec","Type":"ContainerStarted","Data":"79b7f20128890b129d46fc2fadaaa24ce5290fe53aab9f80c2eab6e1e67cbf52"} Oct 07 14:27:25 crc kubenswrapper[4959]: I1007 14:27:25.374916 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-ms8vj" podStartSLOduration=1.8617990660000001 podStartE2EDuration="2.374888211s" podCreationTimestamp="2025-10-07 14:27:23 +0000 UTC" firstStartedPulling="2025-10-07 14:27:24.333527876 +0000 UTC m=+2526.416932201" lastFinishedPulling="2025-10-07 14:27:24.846617021 +0000 UTC m=+2526.930021346" observedRunningTime="2025-10-07 14:27:25.371165897 +0000 UTC m=+2527.454570222" watchObservedRunningTime="2025-10-07 14:27:25.374888211 +0000 UTC m=+2527.458292576" Oct 07 14:27:28 crc kubenswrapper[4959]: I1007 14:27:28.660858 4959 scope.go:117] "RemoveContainer" containerID="0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d" Oct 07 14:27:29 crc kubenswrapper[4959]: I1007 14:27:29.381762 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerStarted","Data":"f740f3b30981ad537e3b0957d254cc4251e26f8688f9d1a5e4dc59cde4189ef5"} Oct 07 14:27:35 crc kubenswrapper[4959]: I1007 14:27:35.440523 4959 generic.go:334] "Generic (PLEG): container finished" podID="634f1665-e210-4c71-9c7a-8cf30b5f3fec" containerID="92b8c54c7449e7c3bfe20512c60b67fff9f168b70ae22f0171766d931d29b264" exitCode=0 Oct 07 14:27:35 crc kubenswrapper[4959]: I1007 14:27:35.440687 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ms8vj" event={"ID":"634f1665-e210-4c71-9c7a-8cf30b5f3fec","Type":"ContainerDied","Data":"92b8c54c7449e7c3bfe20512c60b67fff9f168b70ae22f0171766d931d29b264"} Oct 07 14:27:36 crc kubenswrapper[4959]: I1007 14:27:36.828204 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ms8vj" Oct 07 14:27:36 crc kubenswrapper[4959]: I1007 14:27:36.928136 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/634f1665-e210-4c71-9c7a-8cf30b5f3fec-ceph\") pod \"634f1665-e210-4c71-9c7a-8cf30b5f3fec\" (UID: \"634f1665-e210-4c71-9c7a-8cf30b5f3fec\") " Oct 07 14:27:36 crc kubenswrapper[4959]: I1007 14:27:36.928199 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/634f1665-e210-4c71-9c7a-8cf30b5f3fec-ssh-key-openstack-edpm-ipam\") pod \"634f1665-e210-4c71-9c7a-8cf30b5f3fec\" (UID: \"634f1665-e210-4c71-9c7a-8cf30b5f3fec\") " Oct 07 14:27:36 crc kubenswrapper[4959]: I1007 14:27:36.928318 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6hhg\" (UniqueName: \"kubernetes.io/projected/634f1665-e210-4c71-9c7a-8cf30b5f3fec-kube-api-access-g6hhg\") pod \"634f1665-e210-4c71-9c7a-8cf30b5f3fec\" (UID: \"634f1665-e210-4c71-9c7a-8cf30b5f3fec\") " Oct 07 14:27:36 crc kubenswrapper[4959]: I1007 14:27:36.928339 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/634f1665-e210-4c71-9c7a-8cf30b5f3fec-inventory-0\") pod \"634f1665-e210-4c71-9c7a-8cf30b5f3fec\" (UID: \"634f1665-e210-4c71-9c7a-8cf30b5f3fec\") " Oct 07 14:27:36 crc kubenswrapper[4959]: I1007 14:27:36.934493 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/634f1665-e210-4c71-9c7a-8cf30b5f3fec-ceph" (OuterVolumeSpecName: "ceph") pod "634f1665-e210-4c71-9c7a-8cf30b5f3fec" (UID: "634f1665-e210-4c71-9c7a-8cf30b5f3fec"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:27:36 crc kubenswrapper[4959]: I1007 14:27:36.935453 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/634f1665-e210-4c71-9c7a-8cf30b5f3fec-kube-api-access-g6hhg" (OuterVolumeSpecName: "kube-api-access-g6hhg") pod "634f1665-e210-4c71-9c7a-8cf30b5f3fec" (UID: "634f1665-e210-4c71-9c7a-8cf30b5f3fec"). InnerVolumeSpecName "kube-api-access-g6hhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:27:36 crc kubenswrapper[4959]: I1007 14:27:36.959475 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/634f1665-e210-4c71-9c7a-8cf30b5f3fec-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "634f1665-e210-4c71-9c7a-8cf30b5f3fec" (UID: "634f1665-e210-4c71-9c7a-8cf30b5f3fec"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:27:36 crc kubenswrapper[4959]: I1007 14:27:36.963356 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/634f1665-e210-4c71-9c7a-8cf30b5f3fec-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "634f1665-e210-4c71-9c7a-8cf30b5f3fec" (UID: "634f1665-e210-4c71-9c7a-8cf30b5f3fec"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.029969 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6hhg\" (UniqueName: \"kubernetes.io/projected/634f1665-e210-4c71-9c7a-8cf30b5f3fec-kube-api-access-g6hhg\") on node \"crc\" DevicePath \"\"" Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.030314 4959 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/634f1665-e210-4c71-9c7a-8cf30b5f3fec-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.030325 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/634f1665-e210-4c71-9c7a-8cf30b5f3fec-ceph\") on node \"crc\" DevicePath \"\"" Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.030335 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/634f1665-e210-4c71-9c7a-8cf30b5f3fec-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.457829 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ms8vj" event={"ID":"634f1665-e210-4c71-9c7a-8cf30b5f3fec","Type":"ContainerDied","Data":"79b7f20128890b129d46fc2fadaaa24ce5290fe53aab9f80c2eab6e1e67cbf52"} Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.458173 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79b7f20128890b129d46fc2fadaaa24ce5290fe53aab9f80c2eab6e1e67cbf52" Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.457876 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ms8vj" Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.546496 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-s8dcw"] Oct 07 14:27:37 crc kubenswrapper[4959]: E1007 14:27:37.547070 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="634f1665-e210-4c71-9c7a-8cf30b5f3fec" containerName="ssh-known-hosts-edpm-deployment" Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.547092 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="634f1665-e210-4c71-9c7a-8cf30b5f3fec" containerName="ssh-known-hosts-edpm-deployment" Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.547437 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="634f1665-e210-4c71-9c7a-8cf30b5f3fec" containerName="ssh-known-hosts-edpm-deployment" Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.548329 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s8dcw" Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.554573 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-s8dcw"] Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.555200 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.555319 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.555391 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.555458 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-c6vmh" Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.555206 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.741932 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0e2eb772-e3a8-427f-a226-7aa85dd69238-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-s8dcw\" (UID: \"0e2eb772-e3a8-427f-a226-7aa85dd69238\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s8dcw" Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.741985 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wn6kd\" (UniqueName: \"kubernetes.io/projected/0e2eb772-e3a8-427f-a226-7aa85dd69238-kube-api-access-wn6kd\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-s8dcw\" (UID: \"0e2eb772-e3a8-427f-a226-7aa85dd69238\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s8dcw" Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.742014 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e2eb772-e3a8-427f-a226-7aa85dd69238-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-s8dcw\" (UID: \"0e2eb772-e3a8-427f-a226-7aa85dd69238\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s8dcw" Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.742140 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e2eb772-e3a8-427f-a226-7aa85dd69238-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-s8dcw\" (UID: \"0e2eb772-e3a8-427f-a226-7aa85dd69238\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s8dcw" Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.843875 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wn6kd\" (UniqueName: \"kubernetes.io/projected/0e2eb772-e3a8-427f-a226-7aa85dd69238-kube-api-access-wn6kd\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-s8dcw\" (UID: \"0e2eb772-e3a8-427f-a226-7aa85dd69238\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s8dcw" Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.843928 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e2eb772-e3a8-427f-a226-7aa85dd69238-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-s8dcw\" (UID: \"0e2eb772-e3a8-427f-a226-7aa85dd69238\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s8dcw" Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.844008 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e2eb772-e3a8-427f-a226-7aa85dd69238-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-s8dcw\" (UID: \"0e2eb772-e3a8-427f-a226-7aa85dd69238\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s8dcw" Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.844127 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0e2eb772-e3a8-427f-a226-7aa85dd69238-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-s8dcw\" (UID: \"0e2eb772-e3a8-427f-a226-7aa85dd69238\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s8dcw" Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.849294 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e2eb772-e3a8-427f-a226-7aa85dd69238-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-s8dcw\" (UID: \"0e2eb772-e3a8-427f-a226-7aa85dd69238\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s8dcw" Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.849361 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e2eb772-e3a8-427f-a226-7aa85dd69238-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-s8dcw\" (UID: \"0e2eb772-e3a8-427f-a226-7aa85dd69238\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s8dcw" Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.849894 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0e2eb772-e3a8-427f-a226-7aa85dd69238-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-s8dcw\" (UID: \"0e2eb772-e3a8-427f-a226-7aa85dd69238\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s8dcw" Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.862743 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wn6kd\" (UniqueName: \"kubernetes.io/projected/0e2eb772-e3a8-427f-a226-7aa85dd69238-kube-api-access-wn6kd\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-s8dcw\" (UID: \"0e2eb772-e3a8-427f-a226-7aa85dd69238\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s8dcw" Oct 07 14:27:37 crc kubenswrapper[4959]: I1007 14:27:37.880663 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s8dcw" Oct 07 14:27:38 crc kubenswrapper[4959]: I1007 14:27:38.395561 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-s8dcw"] Oct 07 14:27:38 crc kubenswrapper[4959]: I1007 14:27:38.469435 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s8dcw" event={"ID":"0e2eb772-e3a8-427f-a226-7aa85dd69238","Type":"ContainerStarted","Data":"783b276219042e674ef55a8b10554d3f77a1f387127cd9c06a011167732e4635"} Oct 07 14:27:39 crc kubenswrapper[4959]: I1007 14:27:39.485670 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s8dcw" event={"ID":"0e2eb772-e3a8-427f-a226-7aa85dd69238","Type":"ContainerStarted","Data":"f7e3135ab39e4eee8c9a1033fb465a8f2e03a05fdb2ffedab9daa9842ae1559f"} Oct 07 14:27:39 crc kubenswrapper[4959]: I1007 14:27:39.514536 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s8dcw" podStartSLOduration=1.9239638829999999 podStartE2EDuration="2.514511348s" podCreationTimestamp="2025-10-07 14:27:37 +0000 UTC" firstStartedPulling="2025-10-07 14:27:38.392349128 +0000 UTC m=+2540.475753453" lastFinishedPulling="2025-10-07 14:27:38.982896593 +0000 UTC m=+2541.066300918" observedRunningTime="2025-10-07 14:27:39.502625387 +0000 UTC m=+2541.586029732" watchObservedRunningTime="2025-10-07 14:27:39.514511348 +0000 UTC m=+2541.597915673" Oct 07 14:27:47 crc kubenswrapper[4959]: I1007 14:27:47.549041 4959 generic.go:334] "Generic (PLEG): container finished" podID="0e2eb772-e3a8-427f-a226-7aa85dd69238" containerID="f7e3135ab39e4eee8c9a1033fb465a8f2e03a05fdb2ffedab9daa9842ae1559f" exitCode=0 Oct 07 14:27:47 crc kubenswrapper[4959]: I1007 14:27:47.549136 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s8dcw" event={"ID":"0e2eb772-e3a8-427f-a226-7aa85dd69238","Type":"ContainerDied","Data":"f7e3135ab39e4eee8c9a1033fb465a8f2e03a05fdb2ffedab9daa9842ae1559f"} Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.003843 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s8dcw" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.096535 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wn6kd\" (UniqueName: \"kubernetes.io/projected/0e2eb772-e3a8-427f-a226-7aa85dd69238-kube-api-access-wn6kd\") pod \"0e2eb772-e3a8-427f-a226-7aa85dd69238\" (UID: \"0e2eb772-e3a8-427f-a226-7aa85dd69238\") " Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.096625 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e2eb772-e3a8-427f-a226-7aa85dd69238-inventory\") pod \"0e2eb772-e3a8-427f-a226-7aa85dd69238\" (UID: \"0e2eb772-e3a8-427f-a226-7aa85dd69238\") " Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.096728 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e2eb772-e3a8-427f-a226-7aa85dd69238-ssh-key\") pod \"0e2eb772-e3a8-427f-a226-7aa85dd69238\" (UID: \"0e2eb772-e3a8-427f-a226-7aa85dd69238\") " Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.096767 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0e2eb772-e3a8-427f-a226-7aa85dd69238-ceph\") pod \"0e2eb772-e3a8-427f-a226-7aa85dd69238\" (UID: \"0e2eb772-e3a8-427f-a226-7aa85dd69238\") " Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.103995 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e2eb772-e3a8-427f-a226-7aa85dd69238-kube-api-access-wn6kd" (OuterVolumeSpecName: "kube-api-access-wn6kd") pod "0e2eb772-e3a8-427f-a226-7aa85dd69238" (UID: "0e2eb772-e3a8-427f-a226-7aa85dd69238"). InnerVolumeSpecName "kube-api-access-wn6kd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.108618 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e2eb772-e3a8-427f-a226-7aa85dd69238-ceph" (OuterVolumeSpecName: "ceph") pod "0e2eb772-e3a8-427f-a226-7aa85dd69238" (UID: "0e2eb772-e3a8-427f-a226-7aa85dd69238"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.127762 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e2eb772-e3a8-427f-a226-7aa85dd69238-inventory" (OuterVolumeSpecName: "inventory") pod "0e2eb772-e3a8-427f-a226-7aa85dd69238" (UID: "0e2eb772-e3a8-427f-a226-7aa85dd69238"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.142693 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e2eb772-e3a8-427f-a226-7aa85dd69238-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0e2eb772-e3a8-427f-a226-7aa85dd69238" (UID: "0e2eb772-e3a8-427f-a226-7aa85dd69238"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.200130 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wn6kd\" (UniqueName: \"kubernetes.io/projected/0e2eb772-e3a8-427f-a226-7aa85dd69238-kube-api-access-wn6kd\") on node \"crc\" DevicePath \"\"" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.200167 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e2eb772-e3a8-427f-a226-7aa85dd69238-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.200177 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e2eb772-e3a8-427f-a226-7aa85dd69238-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.200187 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0e2eb772-e3a8-427f-a226-7aa85dd69238-ceph\") on node \"crc\" DevicePath \"\"" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.571009 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s8dcw" event={"ID":"0e2eb772-e3a8-427f-a226-7aa85dd69238","Type":"ContainerDied","Data":"783b276219042e674ef55a8b10554d3f77a1f387127cd9c06a011167732e4635"} Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.571560 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="783b276219042e674ef55a8b10554d3f77a1f387127cd9c06a011167732e4635" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.571060 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-s8dcw" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.646850 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg"] Oct 07 14:27:49 crc kubenswrapper[4959]: E1007 14:27:49.647249 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e2eb772-e3a8-427f-a226-7aa85dd69238" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.647265 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e2eb772-e3a8-427f-a226-7aa85dd69238" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.647462 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e2eb772-e3a8-427f-a226-7aa85dd69238" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.648115 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.652371 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.652381 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.652381 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.656978 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.657869 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-c6vmh" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.660613 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg"] Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.811872 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00ed94a9-b896-493f-b1bd-8e161a45139c-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg\" (UID: \"00ed94a9-b896-493f-b1bd-8e161a45139c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.811945 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00ed94a9-b896-493f-b1bd-8e161a45139c-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg\" (UID: \"00ed94a9-b896-493f-b1bd-8e161a45139c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.812296 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52wk6\" (UniqueName: \"kubernetes.io/projected/00ed94a9-b896-493f-b1bd-8e161a45139c-kube-api-access-52wk6\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg\" (UID: \"00ed94a9-b896-493f-b1bd-8e161a45139c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.812704 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/00ed94a9-b896-493f-b1bd-8e161a45139c-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg\" (UID: \"00ed94a9-b896-493f-b1bd-8e161a45139c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.915990 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00ed94a9-b896-493f-b1bd-8e161a45139c-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg\" (UID: \"00ed94a9-b896-493f-b1bd-8e161a45139c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.916067 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00ed94a9-b896-493f-b1bd-8e161a45139c-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg\" (UID: \"00ed94a9-b896-493f-b1bd-8e161a45139c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.916129 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52wk6\" (UniqueName: \"kubernetes.io/projected/00ed94a9-b896-493f-b1bd-8e161a45139c-kube-api-access-52wk6\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg\" (UID: \"00ed94a9-b896-493f-b1bd-8e161a45139c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.916205 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/00ed94a9-b896-493f-b1bd-8e161a45139c-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg\" (UID: \"00ed94a9-b896-493f-b1bd-8e161a45139c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.922913 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/00ed94a9-b896-493f-b1bd-8e161a45139c-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg\" (UID: \"00ed94a9-b896-493f-b1bd-8e161a45139c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.923431 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00ed94a9-b896-493f-b1bd-8e161a45139c-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg\" (UID: \"00ed94a9-b896-493f-b1bd-8e161a45139c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.925072 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00ed94a9-b896-493f-b1bd-8e161a45139c-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg\" (UID: \"00ed94a9-b896-493f-b1bd-8e161a45139c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.939849 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52wk6\" (UniqueName: \"kubernetes.io/projected/00ed94a9-b896-493f-b1bd-8e161a45139c-kube-api-access-52wk6\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg\" (UID: \"00ed94a9-b896-493f-b1bd-8e161a45139c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg" Oct 07 14:27:49 crc kubenswrapper[4959]: I1007 14:27:49.965571 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg" Oct 07 14:27:50 crc kubenswrapper[4959]: I1007 14:27:50.506075 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg"] Oct 07 14:27:50 crc kubenswrapper[4959]: I1007 14:27:50.592967 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg" event={"ID":"00ed94a9-b896-493f-b1bd-8e161a45139c","Type":"ContainerStarted","Data":"8fc92d4d75376f7670bebe341534c56623cd79b276d4a787840cff0df788e33d"} Oct 07 14:27:51 crc kubenswrapper[4959]: I1007 14:27:51.612984 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg" event={"ID":"00ed94a9-b896-493f-b1bd-8e161a45139c","Type":"ContainerStarted","Data":"de2d2d30b226e08225342430306c9129b4bb7658eb275dfda00e3b2f18db7853"} Oct 07 14:27:51 crc kubenswrapper[4959]: I1007 14:27:51.633377 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg" podStartSLOduration=2.144867897 podStartE2EDuration="2.63334821s" podCreationTimestamp="2025-10-07 14:27:49 +0000 UTC" firstStartedPulling="2025-10-07 14:27:50.506221803 +0000 UTC m=+2552.589626128" lastFinishedPulling="2025-10-07 14:27:50.994702106 +0000 UTC m=+2553.078106441" observedRunningTime="2025-10-07 14:27:51.628812255 +0000 UTC m=+2553.712216580" watchObservedRunningTime="2025-10-07 14:27:51.63334821 +0000 UTC m=+2553.716752535" Oct 07 14:28:01 crc kubenswrapper[4959]: I1007 14:28:01.706464 4959 generic.go:334] "Generic (PLEG): container finished" podID="00ed94a9-b896-493f-b1bd-8e161a45139c" containerID="de2d2d30b226e08225342430306c9129b4bb7658eb275dfda00e3b2f18db7853" exitCode=0 Oct 07 14:28:01 crc kubenswrapper[4959]: I1007 14:28:01.706984 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg" event={"ID":"00ed94a9-b896-493f-b1bd-8e161a45139c","Type":"ContainerDied","Data":"de2d2d30b226e08225342430306c9129b4bb7658eb275dfda00e3b2f18db7853"} Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.202144 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.343225 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00ed94a9-b896-493f-b1bd-8e161a45139c-ssh-key\") pod \"00ed94a9-b896-493f-b1bd-8e161a45139c\" (UID: \"00ed94a9-b896-493f-b1bd-8e161a45139c\") " Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.343415 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00ed94a9-b896-493f-b1bd-8e161a45139c-inventory\") pod \"00ed94a9-b896-493f-b1bd-8e161a45139c\" (UID: \"00ed94a9-b896-493f-b1bd-8e161a45139c\") " Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.343534 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52wk6\" (UniqueName: \"kubernetes.io/projected/00ed94a9-b896-493f-b1bd-8e161a45139c-kube-api-access-52wk6\") pod \"00ed94a9-b896-493f-b1bd-8e161a45139c\" (UID: \"00ed94a9-b896-493f-b1bd-8e161a45139c\") " Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.343711 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/00ed94a9-b896-493f-b1bd-8e161a45139c-ceph\") pod \"00ed94a9-b896-493f-b1bd-8e161a45139c\" (UID: \"00ed94a9-b896-493f-b1bd-8e161a45139c\") " Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.350300 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00ed94a9-b896-493f-b1bd-8e161a45139c-ceph" (OuterVolumeSpecName: "ceph") pod "00ed94a9-b896-493f-b1bd-8e161a45139c" (UID: "00ed94a9-b896-493f-b1bd-8e161a45139c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.355202 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00ed94a9-b896-493f-b1bd-8e161a45139c-kube-api-access-52wk6" (OuterVolumeSpecName: "kube-api-access-52wk6") pod "00ed94a9-b896-493f-b1bd-8e161a45139c" (UID: "00ed94a9-b896-493f-b1bd-8e161a45139c"). InnerVolumeSpecName "kube-api-access-52wk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.375271 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00ed94a9-b896-493f-b1bd-8e161a45139c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "00ed94a9-b896-493f-b1bd-8e161a45139c" (UID: "00ed94a9-b896-493f-b1bd-8e161a45139c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.376469 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00ed94a9-b896-493f-b1bd-8e161a45139c-inventory" (OuterVolumeSpecName: "inventory") pod "00ed94a9-b896-493f-b1bd-8e161a45139c" (UID: "00ed94a9-b896-493f-b1bd-8e161a45139c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.446474 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00ed94a9-b896-493f-b1bd-8e161a45139c-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.446529 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52wk6\" (UniqueName: \"kubernetes.io/projected/00ed94a9-b896-493f-b1bd-8e161a45139c-kube-api-access-52wk6\") on node \"crc\" DevicePath \"\"" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.446544 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/00ed94a9-b896-493f-b1bd-8e161a45139c-ceph\") on node \"crc\" DevicePath \"\"" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.446552 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00ed94a9-b896-493f-b1bd-8e161a45139c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.726030 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg" event={"ID":"00ed94a9-b896-493f-b1bd-8e161a45139c","Type":"ContainerDied","Data":"8fc92d4d75376f7670bebe341534c56623cd79b276d4a787840cff0df788e33d"} Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.726066 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8fc92d4d75376f7670bebe341534c56623cd79b276d4a787840cff0df788e33d" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.726136 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.843264 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v"] Oct 07 14:28:03 crc kubenswrapper[4959]: E1007 14:28:03.844347 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00ed94a9-b896-493f-b1bd-8e161a45139c" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.844477 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="00ed94a9-b896-493f-b1bd-8e161a45139c" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.844820 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="00ed94a9-b896-493f-b1bd-8e161a45139c" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.845875 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.849162 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.849609 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-c6vmh" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.849866 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.850341 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.850500 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.850685 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.850869 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.851160 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.855621 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v"] Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.955526 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dp2xt\" (UniqueName: \"kubernetes.io/projected/d2f04a89-92db-4fae-994a-dc901fa21909-kube-api-access-dp2xt\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.955914 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.956070 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.956146 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.956288 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.956315 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2f04a89-92db-4fae-994a-dc901fa21909-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.956355 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.956397 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2f04a89-92db-4fae-994a-dc901fa21909-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.956465 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.956485 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.956641 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.956727 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:03 crc kubenswrapper[4959]: I1007 14:28:03.956814 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2f04a89-92db-4fae-994a-dc901fa21909-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:04 crc kubenswrapper[4959]: I1007 14:28:04.058418 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dp2xt\" (UniqueName: \"kubernetes.io/projected/d2f04a89-92db-4fae-994a-dc901fa21909-kube-api-access-dp2xt\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:04 crc kubenswrapper[4959]: I1007 14:28:04.058507 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:04 crc kubenswrapper[4959]: I1007 14:28:04.058547 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:04 crc kubenswrapper[4959]: I1007 14:28:04.058578 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:04 crc kubenswrapper[4959]: I1007 14:28:04.058635 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:04 crc kubenswrapper[4959]: I1007 14:28:04.058666 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2f04a89-92db-4fae-994a-dc901fa21909-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:04 crc kubenswrapper[4959]: I1007 14:28:04.058695 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:04 crc kubenswrapper[4959]: I1007 14:28:04.058726 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2f04a89-92db-4fae-994a-dc901fa21909-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:04 crc kubenswrapper[4959]: I1007 14:28:04.058762 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:04 crc kubenswrapper[4959]: I1007 14:28:04.058782 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:04 crc kubenswrapper[4959]: I1007 14:28:04.058839 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:04 crc kubenswrapper[4959]: I1007 14:28:04.058864 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:04 crc kubenswrapper[4959]: I1007 14:28:04.058892 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2f04a89-92db-4fae-994a-dc901fa21909-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:04 crc kubenswrapper[4959]: I1007 14:28:04.062486 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:04 crc kubenswrapper[4959]: I1007 14:28:04.062878 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2f04a89-92db-4fae-994a-dc901fa21909-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:04 crc kubenswrapper[4959]: I1007 14:28:04.063147 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:04 crc kubenswrapper[4959]: I1007 14:28:04.063660 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2f04a89-92db-4fae-994a-dc901fa21909-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:04 crc kubenswrapper[4959]: I1007 14:28:04.063766 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:04 crc kubenswrapper[4959]: I1007 14:28:04.063858 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:04 crc kubenswrapper[4959]: I1007 14:28:04.064006 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:04 crc kubenswrapper[4959]: I1007 14:28:04.065363 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:04 crc kubenswrapper[4959]: I1007 14:28:04.065652 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2f04a89-92db-4fae-994a-dc901fa21909-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:04 crc kubenswrapper[4959]: I1007 14:28:04.066741 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:04 crc kubenswrapper[4959]: I1007 14:28:04.066803 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:04 crc kubenswrapper[4959]: I1007 14:28:04.067791 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:04 crc kubenswrapper[4959]: I1007 14:28:04.076006 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dp2xt\" (UniqueName: \"kubernetes.io/projected/d2f04a89-92db-4fae-994a-dc901fa21909-kube-api-access-dp2xt\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:04 crc kubenswrapper[4959]: I1007 14:28:04.163750 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:04 crc kubenswrapper[4959]: I1007 14:28:04.743202 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v"] Oct 07 14:28:05 crc kubenswrapper[4959]: I1007 14:28:05.743001 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" event={"ID":"d2f04a89-92db-4fae-994a-dc901fa21909","Type":"ContainerStarted","Data":"2ca84728958acaf56fd00c0bf2c610a44becc2962b1239074069358d3f613a95"} Oct 07 14:28:05 crc kubenswrapper[4959]: I1007 14:28:05.743357 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" event={"ID":"d2f04a89-92db-4fae-994a-dc901fa21909","Type":"ContainerStarted","Data":"c6c0a942a676fe9987b415d21be2cdf8b434b84611e999774303ca372b89d944"} Oct 07 14:28:05 crc kubenswrapper[4959]: I1007 14:28:05.765219 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" podStartSLOduration=2.304542582 podStartE2EDuration="2.765196371s" podCreationTimestamp="2025-10-07 14:28:03 +0000 UTC" firstStartedPulling="2025-10-07 14:28:04.774382244 +0000 UTC m=+2566.857786569" lastFinishedPulling="2025-10-07 14:28:05.235036033 +0000 UTC m=+2567.318440358" observedRunningTime="2025-10-07 14:28:05.761769975 +0000 UTC m=+2567.845174300" watchObservedRunningTime="2025-10-07 14:28:05.765196371 +0000 UTC m=+2567.848600696" Oct 07 14:28:39 crc kubenswrapper[4959]: I1007 14:28:39.066048 4959 generic.go:334] "Generic (PLEG): container finished" podID="d2f04a89-92db-4fae-994a-dc901fa21909" containerID="2ca84728958acaf56fd00c0bf2c610a44becc2962b1239074069358d3f613a95" exitCode=0 Oct 07 14:28:39 crc kubenswrapper[4959]: I1007 14:28:39.066124 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" event={"ID":"d2f04a89-92db-4fae-994a-dc901fa21909","Type":"ContainerDied","Data":"2ca84728958acaf56fd00c0bf2c610a44becc2962b1239074069358d3f613a95"} Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.473599 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.649728 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-ovn-combined-ca-bundle\") pod \"d2f04a89-92db-4fae-994a-dc901fa21909\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.649800 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-inventory\") pod \"d2f04a89-92db-4fae-994a-dc901fa21909\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.649846 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-libvirt-combined-ca-bundle\") pod \"d2f04a89-92db-4fae-994a-dc901fa21909\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.649889 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dp2xt\" (UniqueName: \"kubernetes.io/projected/d2f04a89-92db-4fae-994a-dc901fa21909-kube-api-access-dp2xt\") pod \"d2f04a89-92db-4fae-994a-dc901fa21909\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.649912 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-nova-combined-ca-bundle\") pod \"d2f04a89-92db-4fae-994a-dc901fa21909\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.649936 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-ceph\") pod \"d2f04a89-92db-4fae-994a-dc901fa21909\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.649966 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-bootstrap-combined-ca-bundle\") pod \"d2f04a89-92db-4fae-994a-dc901fa21909\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.650041 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2f04a89-92db-4fae-994a-dc901fa21909-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"d2f04a89-92db-4fae-994a-dc901fa21909\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.650077 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-ssh-key\") pod \"d2f04a89-92db-4fae-994a-dc901fa21909\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.650136 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-repo-setup-combined-ca-bundle\") pod \"d2f04a89-92db-4fae-994a-dc901fa21909\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.650164 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2f04a89-92db-4fae-994a-dc901fa21909-openstack-edpm-ipam-ovn-default-certs-0\") pod \"d2f04a89-92db-4fae-994a-dc901fa21909\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.650203 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-neutron-metadata-combined-ca-bundle\") pod \"d2f04a89-92db-4fae-994a-dc901fa21909\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.650275 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2f04a89-92db-4fae-994a-dc901fa21909-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"d2f04a89-92db-4fae-994a-dc901fa21909\" (UID: \"d2f04a89-92db-4fae-994a-dc901fa21909\") " Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.660542 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "d2f04a89-92db-4fae-994a-dc901fa21909" (UID: "d2f04a89-92db-4fae-994a-dc901fa21909"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.660658 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "d2f04a89-92db-4fae-994a-dc901fa21909" (UID: "d2f04a89-92db-4fae-994a-dc901fa21909"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.660799 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2f04a89-92db-4fae-994a-dc901fa21909-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "d2f04a89-92db-4fae-994a-dc901fa21909" (UID: "d2f04a89-92db-4fae-994a-dc901fa21909"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.661056 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2f04a89-92db-4fae-994a-dc901fa21909-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "d2f04a89-92db-4fae-994a-dc901fa21909" (UID: "d2f04a89-92db-4fae-994a-dc901fa21909"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.661138 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "d2f04a89-92db-4fae-994a-dc901fa21909" (UID: "d2f04a89-92db-4fae-994a-dc901fa21909"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.662421 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-ceph" (OuterVolumeSpecName: "ceph") pod "d2f04a89-92db-4fae-994a-dc901fa21909" (UID: "d2f04a89-92db-4fae-994a-dc901fa21909"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.666488 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "d2f04a89-92db-4fae-994a-dc901fa21909" (UID: "d2f04a89-92db-4fae-994a-dc901fa21909"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.666531 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2f04a89-92db-4fae-994a-dc901fa21909-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "d2f04a89-92db-4fae-994a-dc901fa21909" (UID: "d2f04a89-92db-4fae-994a-dc901fa21909"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.668422 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2f04a89-92db-4fae-994a-dc901fa21909-kube-api-access-dp2xt" (OuterVolumeSpecName: "kube-api-access-dp2xt") pod "d2f04a89-92db-4fae-994a-dc901fa21909" (UID: "d2f04a89-92db-4fae-994a-dc901fa21909"). InnerVolumeSpecName "kube-api-access-dp2xt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.670325 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "d2f04a89-92db-4fae-994a-dc901fa21909" (UID: "d2f04a89-92db-4fae-994a-dc901fa21909"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.674087 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "d2f04a89-92db-4fae-994a-dc901fa21909" (UID: "d2f04a89-92db-4fae-994a-dc901fa21909"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.686158 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-inventory" (OuterVolumeSpecName: "inventory") pod "d2f04a89-92db-4fae-994a-dc901fa21909" (UID: "d2f04a89-92db-4fae-994a-dc901fa21909"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.698392 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d2f04a89-92db-4fae-994a-dc901fa21909" (UID: "d2f04a89-92db-4fae-994a-dc901fa21909"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.752456 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.752498 4959 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2f04a89-92db-4fae-994a-dc901fa21909-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.752527 4959 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.752548 4959 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2f04a89-92db-4fae-994a-dc901fa21909-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.752562 4959 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.752576 4959 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2f04a89-92db-4fae-994a-dc901fa21909-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.752591 4959 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.752603 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.752612 4959 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.752622 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dp2xt\" (UniqueName: \"kubernetes.io/projected/d2f04a89-92db-4fae-994a-dc901fa21909-kube-api-access-dp2xt\") on node \"crc\" DevicePath \"\"" Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.752630 4959 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.752639 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-ceph\") on node \"crc\" DevicePath \"\"" Oct 07 14:28:40 crc kubenswrapper[4959]: I1007 14:28:40.752647 4959 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f04a89-92db-4fae-994a-dc901fa21909-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:28:41 crc kubenswrapper[4959]: I1007 14:28:41.087015 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" event={"ID":"d2f04a89-92db-4fae-994a-dc901fa21909","Type":"ContainerDied","Data":"c6c0a942a676fe9987b415d21be2cdf8b434b84611e999774303ca372b89d944"} Oct 07 14:28:41 crc kubenswrapper[4959]: I1007 14:28:41.087071 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6c0a942a676fe9987b415d21be2cdf8b434b84611e999774303ca372b89d944" Oct 07 14:28:41 crc kubenswrapper[4959]: I1007 14:28:41.087083 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v" Oct 07 14:28:41 crc kubenswrapper[4959]: I1007 14:28:41.209935 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl"] Oct 07 14:28:41 crc kubenswrapper[4959]: E1007 14:28:41.210605 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2f04a89-92db-4fae-994a-dc901fa21909" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 07 14:28:41 crc kubenswrapper[4959]: I1007 14:28:41.210626 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2f04a89-92db-4fae-994a-dc901fa21909" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 07 14:28:41 crc kubenswrapper[4959]: I1007 14:28:41.210856 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2f04a89-92db-4fae-994a-dc901fa21909" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 07 14:28:41 crc kubenswrapper[4959]: I1007 14:28:41.211930 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl" Oct 07 14:28:41 crc kubenswrapper[4959]: I1007 14:28:41.214745 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 14:28:41 crc kubenswrapper[4959]: I1007 14:28:41.215424 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-c6vmh" Oct 07 14:28:41 crc kubenswrapper[4959]: I1007 14:28:41.215518 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 14:28:41 crc kubenswrapper[4959]: I1007 14:28:41.215907 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 07 14:28:41 crc kubenswrapper[4959]: I1007 14:28:41.216224 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 14:28:41 crc kubenswrapper[4959]: I1007 14:28:41.222595 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl"] Oct 07 14:28:41 crc kubenswrapper[4959]: I1007 14:28:41.366753 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/defafb0d-5ea5-40f6-9918-4c97a946219c-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl\" (UID: \"defafb0d-5ea5-40f6-9918-4c97a946219c\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl" Oct 07 14:28:41 crc kubenswrapper[4959]: I1007 14:28:41.366893 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/defafb0d-5ea5-40f6-9918-4c97a946219c-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl\" (UID: \"defafb0d-5ea5-40f6-9918-4c97a946219c\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl" Oct 07 14:28:41 crc kubenswrapper[4959]: I1007 14:28:41.366932 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w4wf\" (UniqueName: \"kubernetes.io/projected/defafb0d-5ea5-40f6-9918-4c97a946219c-kube-api-access-8w4wf\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl\" (UID: \"defafb0d-5ea5-40f6-9918-4c97a946219c\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl" Oct 07 14:28:41 crc kubenswrapper[4959]: I1007 14:28:41.366969 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/defafb0d-5ea5-40f6-9918-4c97a946219c-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl\" (UID: \"defafb0d-5ea5-40f6-9918-4c97a946219c\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl" Oct 07 14:28:41 crc kubenswrapper[4959]: I1007 14:28:41.468859 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/defafb0d-5ea5-40f6-9918-4c97a946219c-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl\" (UID: \"defafb0d-5ea5-40f6-9918-4c97a946219c\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl" Oct 07 14:28:41 crc kubenswrapper[4959]: I1007 14:28:41.469233 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w4wf\" (UniqueName: \"kubernetes.io/projected/defafb0d-5ea5-40f6-9918-4c97a946219c-kube-api-access-8w4wf\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl\" (UID: \"defafb0d-5ea5-40f6-9918-4c97a946219c\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl" Oct 07 14:28:41 crc kubenswrapper[4959]: I1007 14:28:41.469273 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/defafb0d-5ea5-40f6-9918-4c97a946219c-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl\" (UID: \"defafb0d-5ea5-40f6-9918-4c97a946219c\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl" Oct 07 14:28:41 crc kubenswrapper[4959]: I1007 14:28:41.469326 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/defafb0d-5ea5-40f6-9918-4c97a946219c-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl\" (UID: \"defafb0d-5ea5-40f6-9918-4c97a946219c\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl" Oct 07 14:28:41 crc kubenswrapper[4959]: I1007 14:28:41.473351 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/defafb0d-5ea5-40f6-9918-4c97a946219c-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl\" (UID: \"defafb0d-5ea5-40f6-9918-4c97a946219c\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl" Oct 07 14:28:41 crc kubenswrapper[4959]: I1007 14:28:41.473411 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/defafb0d-5ea5-40f6-9918-4c97a946219c-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl\" (UID: \"defafb0d-5ea5-40f6-9918-4c97a946219c\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl" Oct 07 14:28:41 crc kubenswrapper[4959]: I1007 14:28:41.474383 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/defafb0d-5ea5-40f6-9918-4c97a946219c-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl\" (UID: \"defafb0d-5ea5-40f6-9918-4c97a946219c\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl" Oct 07 14:28:41 crc kubenswrapper[4959]: I1007 14:28:41.489533 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w4wf\" (UniqueName: \"kubernetes.io/projected/defafb0d-5ea5-40f6-9918-4c97a946219c-kube-api-access-8w4wf\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl\" (UID: \"defafb0d-5ea5-40f6-9918-4c97a946219c\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl" Oct 07 14:28:41 crc kubenswrapper[4959]: I1007 14:28:41.544722 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl" Oct 07 14:28:42 crc kubenswrapper[4959]: I1007 14:28:42.080302 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl"] Oct 07 14:28:42 crc kubenswrapper[4959]: I1007 14:28:42.096842 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl" event={"ID":"defafb0d-5ea5-40f6-9918-4c97a946219c","Type":"ContainerStarted","Data":"7f52bd47c9ff1af31ef8ca60a231e0b2301438c6c6dfde40bf92b7fae908470d"} Oct 07 14:28:43 crc kubenswrapper[4959]: I1007 14:28:43.105497 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl" event={"ID":"defafb0d-5ea5-40f6-9918-4c97a946219c","Type":"ContainerStarted","Data":"2d0bc18fb0ccc05f109c4375d48fa338a840f87e48bbcd4bc76367b1abaa963f"} Oct 07 14:28:43 crc kubenswrapper[4959]: I1007 14:28:43.124915 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl" podStartSLOduration=1.584202785 podStartE2EDuration="2.124893889s" podCreationTimestamp="2025-10-07 14:28:41 +0000 UTC" firstStartedPulling="2025-10-07 14:28:42.085041022 +0000 UTC m=+2604.168445347" lastFinishedPulling="2025-10-07 14:28:42.625732126 +0000 UTC m=+2604.709136451" observedRunningTime="2025-10-07 14:28:43.119948664 +0000 UTC m=+2605.203352999" watchObservedRunningTime="2025-10-07 14:28:43.124893889 +0000 UTC m=+2605.208298214" Oct 07 14:28:49 crc kubenswrapper[4959]: I1007 14:28:49.156322 4959 generic.go:334] "Generic (PLEG): container finished" podID="defafb0d-5ea5-40f6-9918-4c97a946219c" containerID="2d0bc18fb0ccc05f109c4375d48fa338a840f87e48bbcd4bc76367b1abaa963f" exitCode=0 Oct 07 14:28:49 crc kubenswrapper[4959]: I1007 14:28:49.156371 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl" event={"ID":"defafb0d-5ea5-40f6-9918-4c97a946219c","Type":"ContainerDied","Data":"2d0bc18fb0ccc05f109c4375d48fa338a840f87e48bbcd4bc76367b1abaa963f"} Oct 07 14:28:50 crc kubenswrapper[4959]: I1007 14:28:50.638221 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl" Oct 07 14:28:50 crc kubenswrapper[4959]: I1007 14:28:50.763324 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/defafb0d-5ea5-40f6-9918-4c97a946219c-ceph\") pod \"defafb0d-5ea5-40f6-9918-4c97a946219c\" (UID: \"defafb0d-5ea5-40f6-9918-4c97a946219c\") " Oct 07 14:28:50 crc kubenswrapper[4959]: I1007 14:28:50.763435 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/defafb0d-5ea5-40f6-9918-4c97a946219c-ssh-key\") pod \"defafb0d-5ea5-40f6-9918-4c97a946219c\" (UID: \"defafb0d-5ea5-40f6-9918-4c97a946219c\") " Oct 07 14:28:50 crc kubenswrapper[4959]: I1007 14:28:50.763490 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/defafb0d-5ea5-40f6-9918-4c97a946219c-inventory\") pod \"defafb0d-5ea5-40f6-9918-4c97a946219c\" (UID: \"defafb0d-5ea5-40f6-9918-4c97a946219c\") " Oct 07 14:28:50 crc kubenswrapper[4959]: I1007 14:28:50.763523 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8w4wf\" (UniqueName: \"kubernetes.io/projected/defafb0d-5ea5-40f6-9918-4c97a946219c-kube-api-access-8w4wf\") pod \"defafb0d-5ea5-40f6-9918-4c97a946219c\" (UID: \"defafb0d-5ea5-40f6-9918-4c97a946219c\") " Oct 07 14:28:50 crc kubenswrapper[4959]: I1007 14:28:50.770316 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/defafb0d-5ea5-40f6-9918-4c97a946219c-kube-api-access-8w4wf" (OuterVolumeSpecName: "kube-api-access-8w4wf") pod "defafb0d-5ea5-40f6-9918-4c97a946219c" (UID: "defafb0d-5ea5-40f6-9918-4c97a946219c"). InnerVolumeSpecName "kube-api-access-8w4wf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:28:50 crc kubenswrapper[4959]: I1007 14:28:50.772225 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/defafb0d-5ea5-40f6-9918-4c97a946219c-ceph" (OuterVolumeSpecName: "ceph") pod "defafb0d-5ea5-40f6-9918-4c97a946219c" (UID: "defafb0d-5ea5-40f6-9918-4c97a946219c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:28:50 crc kubenswrapper[4959]: I1007 14:28:50.789776 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/defafb0d-5ea5-40f6-9918-4c97a946219c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "defafb0d-5ea5-40f6-9918-4c97a946219c" (UID: "defafb0d-5ea5-40f6-9918-4c97a946219c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:28:50 crc kubenswrapper[4959]: I1007 14:28:50.817637 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/defafb0d-5ea5-40f6-9918-4c97a946219c-inventory" (OuterVolumeSpecName: "inventory") pod "defafb0d-5ea5-40f6-9918-4c97a946219c" (UID: "defafb0d-5ea5-40f6-9918-4c97a946219c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:28:50 crc kubenswrapper[4959]: I1007 14:28:50.866523 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/defafb0d-5ea5-40f6-9918-4c97a946219c-ceph\") on node \"crc\" DevicePath \"\"" Oct 07 14:28:50 crc kubenswrapper[4959]: I1007 14:28:50.866795 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/defafb0d-5ea5-40f6-9918-4c97a946219c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 14:28:50 crc kubenswrapper[4959]: I1007 14:28:50.866876 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/defafb0d-5ea5-40f6-9918-4c97a946219c-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 14:28:50 crc kubenswrapper[4959]: I1007 14:28:50.866955 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8w4wf\" (UniqueName: \"kubernetes.io/projected/defafb0d-5ea5-40f6-9918-4c97a946219c-kube-api-access-8w4wf\") on node \"crc\" DevicePath \"\"" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.176896 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl" event={"ID":"defafb0d-5ea5-40f6-9918-4c97a946219c","Type":"ContainerDied","Data":"7f52bd47c9ff1af31ef8ca60a231e0b2301438c6c6dfde40bf92b7fae908470d"} Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.176937 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f52bd47c9ff1af31ef8ca60a231e0b2301438c6c6dfde40bf92b7fae908470d" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.176954 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.238406 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n"] Oct 07 14:28:51 crc kubenswrapper[4959]: E1007 14:28:51.238850 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="defafb0d-5ea5-40f6-9918-4c97a946219c" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.238864 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="defafb0d-5ea5-40f6-9918-4c97a946219c" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.239068 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="defafb0d-5ea5-40f6-9918-4c97a946219c" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.244509 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.246471 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.246752 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.246871 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.246994 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-c6vmh" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.247218 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.247323 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.248260 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n"] Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.376936 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lt62n\" (UID: \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.377013 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lt62n\" (UID: \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.377093 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lt62n\" (UID: \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.377164 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lt62n\" (UID: \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.377187 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gk4pq\" (UniqueName: \"kubernetes.io/projected/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-kube-api-access-gk4pq\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lt62n\" (UID: \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.377285 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lt62n\" (UID: \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.478764 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lt62n\" (UID: \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.478808 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lt62n\" (UID: \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.478833 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lt62n\" (UID: \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.478883 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lt62n\" (UID: \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.478907 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gk4pq\" (UniqueName: \"kubernetes.io/projected/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-kube-api-access-gk4pq\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lt62n\" (UID: \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.478983 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lt62n\" (UID: \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.479826 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lt62n\" (UID: \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.483782 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lt62n\" (UID: \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.483782 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lt62n\" (UID: \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.483861 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lt62n\" (UID: \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.488269 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lt62n\" (UID: \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.495890 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gk4pq\" (UniqueName: \"kubernetes.io/projected/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-kube-api-access-gk4pq\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lt62n\" (UID: \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n" Oct 07 14:28:51 crc kubenswrapper[4959]: I1007 14:28:51.564348 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n" Oct 07 14:28:52 crc kubenswrapper[4959]: I1007 14:28:52.073028 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n"] Oct 07 14:28:52 crc kubenswrapper[4959]: I1007 14:28:52.189046 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n" event={"ID":"64d9894c-3fa2-4dfb-85c7-c8596268ff6b","Type":"ContainerStarted","Data":"b2693479c10ae5006d4325884ca2b5d784acdaaece31b319c24db98302b27a51"} Oct 07 14:28:53 crc kubenswrapper[4959]: I1007 14:28:53.198410 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n" event={"ID":"64d9894c-3fa2-4dfb-85c7-c8596268ff6b","Type":"ContainerStarted","Data":"46e5754d09862f2f4d62b64183acf477f4d944374e7b8645ab0ecad50e8eb717"} Oct 07 14:28:53 crc kubenswrapper[4959]: I1007 14:28:53.226902 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n" podStartSLOduration=1.6042548490000001 podStartE2EDuration="2.226879977s" podCreationTimestamp="2025-10-07 14:28:51 +0000 UTC" firstStartedPulling="2025-10-07 14:28:52.082138035 +0000 UTC m=+2614.165542360" lastFinishedPulling="2025-10-07 14:28:52.704763163 +0000 UTC m=+2614.788167488" observedRunningTime="2025-10-07 14:28:53.218116755 +0000 UTC m=+2615.301521090" watchObservedRunningTime="2025-10-07 14:28:53.226879977 +0000 UTC m=+2615.310284312" Oct 07 14:29:50 crc kubenswrapper[4959]: I1007 14:29:50.629868 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:29:50 crc kubenswrapper[4959]: I1007 14:29:50.630451 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:30:00 crc kubenswrapper[4959]: I1007 14:30:00.141918 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330790-nrtvc"] Oct 07 14:30:00 crc kubenswrapper[4959]: I1007 14:30:00.144002 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330790-nrtvc" Oct 07 14:30:00 crc kubenswrapper[4959]: I1007 14:30:00.146387 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 07 14:30:00 crc kubenswrapper[4959]: I1007 14:30:00.156235 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330790-nrtvc"] Oct 07 14:30:00 crc kubenswrapper[4959]: I1007 14:30:00.156996 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 07 14:30:00 crc kubenswrapper[4959]: I1007 14:30:00.233313 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab-secret-volume\") pod \"collect-profiles-29330790-nrtvc\" (UID: \"2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330790-nrtvc" Oct 07 14:30:00 crc kubenswrapper[4959]: I1007 14:30:00.233429 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jfgs\" (UniqueName: \"kubernetes.io/projected/2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab-kube-api-access-7jfgs\") pod \"collect-profiles-29330790-nrtvc\" (UID: \"2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330790-nrtvc" Oct 07 14:30:00 crc kubenswrapper[4959]: I1007 14:30:00.233457 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab-config-volume\") pod \"collect-profiles-29330790-nrtvc\" (UID: \"2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330790-nrtvc" Oct 07 14:30:00 crc kubenswrapper[4959]: I1007 14:30:00.335447 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab-secret-volume\") pod \"collect-profiles-29330790-nrtvc\" (UID: \"2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330790-nrtvc" Oct 07 14:30:00 crc kubenswrapper[4959]: I1007 14:30:00.335573 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jfgs\" (UniqueName: \"kubernetes.io/projected/2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab-kube-api-access-7jfgs\") pod \"collect-profiles-29330790-nrtvc\" (UID: \"2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330790-nrtvc" Oct 07 14:30:00 crc kubenswrapper[4959]: I1007 14:30:00.335601 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab-config-volume\") pod \"collect-profiles-29330790-nrtvc\" (UID: \"2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330790-nrtvc" Oct 07 14:30:00 crc kubenswrapper[4959]: I1007 14:30:00.336738 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab-config-volume\") pod \"collect-profiles-29330790-nrtvc\" (UID: \"2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330790-nrtvc" Oct 07 14:30:00 crc kubenswrapper[4959]: I1007 14:30:00.350416 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab-secret-volume\") pod \"collect-profiles-29330790-nrtvc\" (UID: \"2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330790-nrtvc" Oct 07 14:30:00 crc kubenswrapper[4959]: I1007 14:30:00.355435 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jfgs\" (UniqueName: \"kubernetes.io/projected/2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab-kube-api-access-7jfgs\") pod \"collect-profiles-29330790-nrtvc\" (UID: \"2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330790-nrtvc" Oct 07 14:30:00 crc kubenswrapper[4959]: I1007 14:30:00.473912 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330790-nrtvc" Oct 07 14:30:00 crc kubenswrapper[4959]: I1007 14:30:00.915757 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330790-nrtvc"] Oct 07 14:30:01 crc kubenswrapper[4959]: I1007 14:30:01.869064 4959 generic.go:334] "Generic (PLEG): container finished" podID="2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab" containerID="766f0cd77ccad16e5276ddfc64d80202ceae10e653ab4a0fbcdd693e1a845aa1" exitCode=0 Oct 07 14:30:01 crc kubenswrapper[4959]: I1007 14:30:01.869384 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330790-nrtvc" event={"ID":"2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab","Type":"ContainerDied","Data":"766f0cd77ccad16e5276ddfc64d80202ceae10e653ab4a0fbcdd693e1a845aa1"} Oct 07 14:30:01 crc kubenswrapper[4959]: I1007 14:30:01.872424 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330790-nrtvc" event={"ID":"2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab","Type":"ContainerStarted","Data":"a5a4d9749bf92836d84a477705811296857466121c09efb7c9e8bb74acce5326"} Oct 07 14:30:03 crc kubenswrapper[4959]: I1007 14:30:03.212791 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330790-nrtvc" Oct 07 14:30:03 crc kubenswrapper[4959]: I1007 14:30:03.301730 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jfgs\" (UniqueName: \"kubernetes.io/projected/2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab-kube-api-access-7jfgs\") pod \"2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab\" (UID: \"2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab\") " Oct 07 14:30:03 crc kubenswrapper[4959]: I1007 14:30:03.301844 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab-secret-volume\") pod \"2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab\" (UID: \"2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab\") " Oct 07 14:30:03 crc kubenswrapper[4959]: I1007 14:30:03.301980 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab-config-volume\") pod \"2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab\" (UID: \"2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab\") " Oct 07 14:30:03 crc kubenswrapper[4959]: I1007 14:30:03.302725 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab-config-volume" (OuterVolumeSpecName: "config-volume") pod "2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab" (UID: "2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:30:03 crc kubenswrapper[4959]: I1007 14:30:03.308832 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab" (UID: "2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:30:03 crc kubenswrapper[4959]: I1007 14:30:03.308907 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab-kube-api-access-7jfgs" (OuterVolumeSpecName: "kube-api-access-7jfgs") pod "2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab" (UID: "2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab"). InnerVolumeSpecName "kube-api-access-7jfgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:30:03 crc kubenswrapper[4959]: I1007 14:30:03.403901 4959 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 14:30:03 crc kubenswrapper[4959]: I1007 14:30:03.403949 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jfgs\" (UniqueName: \"kubernetes.io/projected/2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab-kube-api-access-7jfgs\") on node \"crc\" DevicePath \"\"" Oct 07 14:30:03 crc kubenswrapper[4959]: I1007 14:30:03.403966 4959 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 07 14:30:03 crc kubenswrapper[4959]: I1007 14:30:03.889493 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330790-nrtvc" event={"ID":"2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab","Type":"ContainerDied","Data":"a5a4d9749bf92836d84a477705811296857466121c09efb7c9e8bb74acce5326"} Oct 07 14:30:03 crc kubenswrapper[4959]: I1007 14:30:03.889546 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5a4d9749bf92836d84a477705811296857466121c09efb7c9e8bb74acce5326" Oct 07 14:30:03 crc kubenswrapper[4959]: I1007 14:30:03.889613 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330790-nrtvc" Oct 07 14:30:04 crc kubenswrapper[4959]: I1007 14:30:04.286129 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330745-wbwkx"] Oct 07 14:30:04 crc kubenswrapper[4959]: I1007 14:30:04.296036 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330745-wbwkx"] Oct 07 14:30:04 crc kubenswrapper[4959]: I1007 14:30:04.675423 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0a596b0-4162-4b10-8880-3ea0421b8c74" path="/var/lib/kubelet/pods/d0a596b0-4162-4b10-8880-3ea0421b8c74/volumes" Oct 07 14:30:07 crc kubenswrapper[4959]: I1007 14:30:07.928049 4959 generic.go:334] "Generic (PLEG): container finished" podID="64d9894c-3fa2-4dfb-85c7-c8596268ff6b" containerID="46e5754d09862f2f4d62b64183acf477f4d944374e7b8645ab0ecad50e8eb717" exitCode=0 Oct 07 14:30:07 crc kubenswrapper[4959]: I1007 14:30:07.928152 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n" event={"ID":"64d9894c-3fa2-4dfb-85c7-c8596268ff6b","Type":"ContainerDied","Data":"46e5754d09862f2f4d62b64183acf477f4d944374e7b8645ab0ecad50e8eb717"} Oct 07 14:30:09 crc kubenswrapper[4959]: I1007 14:30:09.403256 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n" Oct 07 14:30:09 crc kubenswrapper[4959]: I1007 14:30:09.536755 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-ssh-key\") pod \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\" (UID: \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\") " Oct 07 14:30:09 crc kubenswrapper[4959]: I1007 14:30:09.536892 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-ceph\") pod \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\" (UID: \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\") " Oct 07 14:30:09 crc kubenswrapper[4959]: I1007 14:30:09.536932 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-inventory\") pod \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\" (UID: \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\") " Oct 07 14:30:09 crc kubenswrapper[4959]: I1007 14:30:09.537131 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gk4pq\" (UniqueName: \"kubernetes.io/projected/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-kube-api-access-gk4pq\") pod \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\" (UID: \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\") " Oct 07 14:30:09 crc kubenswrapper[4959]: I1007 14:30:09.537366 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-ovncontroller-config-0\") pod \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\" (UID: \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\") " Oct 07 14:30:09 crc kubenswrapper[4959]: I1007 14:30:09.537425 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-ovn-combined-ca-bundle\") pod \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\" (UID: \"64d9894c-3fa2-4dfb-85c7-c8596268ff6b\") " Oct 07 14:30:09 crc kubenswrapper[4959]: I1007 14:30:09.545817 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "64d9894c-3fa2-4dfb-85c7-c8596268ff6b" (UID: "64d9894c-3fa2-4dfb-85c7-c8596268ff6b"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:30:09 crc kubenswrapper[4959]: I1007 14:30:09.545954 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-ceph" (OuterVolumeSpecName: "ceph") pod "64d9894c-3fa2-4dfb-85c7-c8596268ff6b" (UID: "64d9894c-3fa2-4dfb-85c7-c8596268ff6b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:30:09 crc kubenswrapper[4959]: I1007 14:30:09.550388 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-kube-api-access-gk4pq" (OuterVolumeSpecName: "kube-api-access-gk4pq") pod "64d9894c-3fa2-4dfb-85c7-c8596268ff6b" (UID: "64d9894c-3fa2-4dfb-85c7-c8596268ff6b"). InnerVolumeSpecName "kube-api-access-gk4pq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:30:09 crc kubenswrapper[4959]: I1007 14:30:09.568713 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "64d9894c-3fa2-4dfb-85c7-c8596268ff6b" (UID: "64d9894c-3fa2-4dfb-85c7-c8596268ff6b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:30:09 crc kubenswrapper[4959]: I1007 14:30:09.573355 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-inventory" (OuterVolumeSpecName: "inventory") pod "64d9894c-3fa2-4dfb-85c7-c8596268ff6b" (UID: "64d9894c-3fa2-4dfb-85c7-c8596268ff6b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:30:09 crc kubenswrapper[4959]: I1007 14:30:09.587059 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "64d9894c-3fa2-4dfb-85c7-c8596268ff6b" (UID: "64d9894c-3fa2-4dfb-85c7-c8596268ff6b"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:30:09 crc kubenswrapper[4959]: I1007 14:30:09.639757 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 14:30:09 crc kubenswrapper[4959]: I1007 14:30:09.639795 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 14:30:09 crc kubenswrapper[4959]: I1007 14:30:09.639810 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-ceph\") on node \"crc\" DevicePath \"\"" Oct 07 14:30:09 crc kubenswrapper[4959]: I1007 14:30:09.639818 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gk4pq\" (UniqueName: \"kubernetes.io/projected/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-kube-api-access-gk4pq\") on node \"crc\" DevicePath \"\"" Oct 07 14:30:09 crc kubenswrapper[4959]: I1007 14:30:09.639832 4959 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 07 14:30:09 crc kubenswrapper[4959]: I1007 14:30:09.639843 4959 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64d9894c-3fa2-4dfb-85c7-c8596268ff6b-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:30:09 crc kubenswrapper[4959]: I1007 14:30:09.954779 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n" event={"ID":"64d9894c-3fa2-4dfb-85c7-c8596268ff6b","Type":"ContainerDied","Data":"b2693479c10ae5006d4325884ca2b5d784acdaaece31b319c24db98302b27a51"} Oct 07 14:30:09 crc kubenswrapper[4959]: I1007 14:30:09.954852 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lt62n" Oct 07 14:30:09 crc kubenswrapper[4959]: I1007 14:30:09.954923 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2693479c10ae5006d4325884ca2b5d784acdaaece31b319c24db98302b27a51" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.178807 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb"] Oct 07 14:30:10 crc kubenswrapper[4959]: E1007 14:30:10.179507 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab" containerName="collect-profiles" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.179525 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab" containerName="collect-profiles" Oct 07 14:30:10 crc kubenswrapper[4959]: E1007 14:30:10.179551 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64d9894c-3fa2-4dfb-85c7-c8596268ff6b" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.179561 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="64d9894c-3fa2-4dfb-85c7-c8596268ff6b" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.179799 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab" containerName="collect-profiles" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.179824 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="64d9894c-3fa2-4dfb-85c7-c8596268ff6b" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.180642 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.184022 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.184220 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.184355 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.184470 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.184505 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.184519 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.184598 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-c6vmh" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.191353 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb"] Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.253292 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb\" (UID: \"6645e33c-3472-41cd-8f39-a702210bc860\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.253370 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlzbz\" (UniqueName: \"kubernetes.io/projected/6645e33c-3472-41cd-8f39-a702210bc860-kube-api-access-zlzbz\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb\" (UID: \"6645e33c-3472-41cd-8f39-a702210bc860\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.253433 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb\" (UID: \"6645e33c-3472-41cd-8f39-a702210bc860\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.253543 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb\" (UID: \"6645e33c-3472-41cd-8f39-a702210bc860\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.253618 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb\" (UID: \"6645e33c-3472-41cd-8f39-a702210bc860\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.253693 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb\" (UID: \"6645e33c-3472-41cd-8f39-a702210bc860\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.253744 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb\" (UID: \"6645e33c-3472-41cd-8f39-a702210bc860\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.355389 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb\" (UID: \"6645e33c-3472-41cd-8f39-a702210bc860\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.355451 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb\" (UID: \"6645e33c-3472-41cd-8f39-a702210bc860\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.355516 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb\" (UID: \"6645e33c-3472-41cd-8f39-a702210bc860\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.355540 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb\" (UID: \"6645e33c-3472-41cd-8f39-a702210bc860\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.355629 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb\" (UID: \"6645e33c-3472-41cd-8f39-a702210bc860\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.355656 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlzbz\" (UniqueName: \"kubernetes.io/projected/6645e33c-3472-41cd-8f39-a702210bc860-kube-api-access-zlzbz\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb\" (UID: \"6645e33c-3472-41cd-8f39-a702210bc860\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.355692 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb\" (UID: \"6645e33c-3472-41cd-8f39-a702210bc860\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.360553 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb\" (UID: \"6645e33c-3472-41cd-8f39-a702210bc860\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.360863 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb\" (UID: \"6645e33c-3472-41cd-8f39-a702210bc860\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.361023 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb\" (UID: \"6645e33c-3472-41cd-8f39-a702210bc860\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.361611 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb\" (UID: \"6645e33c-3472-41cd-8f39-a702210bc860\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.361683 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb\" (UID: \"6645e33c-3472-41cd-8f39-a702210bc860\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.361941 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb\" (UID: \"6645e33c-3472-41cd-8f39-a702210bc860\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.375367 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlzbz\" (UniqueName: \"kubernetes.io/projected/6645e33c-3472-41cd-8f39-a702210bc860-kube-api-access-zlzbz\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb\" (UID: \"6645e33c-3472-41cd-8f39-a702210bc860\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" Oct 07 14:30:10 crc kubenswrapper[4959]: I1007 14:30:10.508990 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" Oct 07 14:30:11 crc kubenswrapper[4959]: I1007 14:30:11.088439 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb"] Oct 07 14:30:11 crc kubenswrapper[4959]: W1007 14:30:11.099656 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6645e33c_3472_41cd_8f39_a702210bc860.slice/crio-c523f365cdcb39573aacc4d986a6233aee1d7ad90df684a98e291b41914a9d60 WatchSource:0}: Error finding container c523f365cdcb39573aacc4d986a6233aee1d7ad90df684a98e291b41914a9d60: Status 404 returned error can't find the container with id c523f365cdcb39573aacc4d986a6233aee1d7ad90df684a98e291b41914a9d60 Oct 07 14:30:11 crc kubenswrapper[4959]: I1007 14:30:11.981544 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" event={"ID":"6645e33c-3472-41cd-8f39-a702210bc860","Type":"ContainerStarted","Data":"b9ed1f1b16adb671f43f9a2773821695ee0a65f46ef494ac485b899221601ef7"} Oct 07 14:30:11 crc kubenswrapper[4959]: I1007 14:30:11.982295 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" event={"ID":"6645e33c-3472-41cd-8f39-a702210bc860","Type":"ContainerStarted","Data":"c523f365cdcb39573aacc4d986a6233aee1d7ad90df684a98e291b41914a9d60"} Oct 07 14:30:12 crc kubenswrapper[4959]: I1007 14:30:12.011831 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" podStartSLOduration=1.529131246 podStartE2EDuration="2.011808883s" podCreationTimestamp="2025-10-07 14:30:10 +0000 UTC" firstStartedPulling="2025-10-07 14:30:11.102749925 +0000 UTC m=+2693.186154250" lastFinishedPulling="2025-10-07 14:30:11.585427562 +0000 UTC m=+2693.668831887" observedRunningTime="2025-10-07 14:30:12.010848489 +0000 UTC m=+2694.094252824" watchObservedRunningTime="2025-10-07 14:30:12.011808883 +0000 UTC m=+2694.095213208" Oct 07 14:30:20 crc kubenswrapper[4959]: I1007 14:30:20.629798 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:30:20 crc kubenswrapper[4959]: I1007 14:30:20.630369 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:30:30 crc kubenswrapper[4959]: I1007 14:30:30.344792 4959 scope.go:117] "RemoveContainer" containerID="a7017d01f13370140f49dfc5cb484478c43956cb32641304263fe6937e7d534b" Oct 07 14:30:50 crc kubenswrapper[4959]: I1007 14:30:50.630835 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:30:50 crc kubenswrapper[4959]: I1007 14:30:50.631948 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:30:50 crc kubenswrapper[4959]: I1007 14:30:50.632041 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 14:30:50 crc kubenswrapper[4959]: I1007 14:30:50.633372 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f740f3b30981ad537e3b0957d254cc4251e26f8688f9d1a5e4dc59cde4189ef5"} pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 14:30:50 crc kubenswrapper[4959]: I1007 14:30:50.633466 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" containerID="cri-o://f740f3b30981ad537e3b0957d254cc4251e26f8688f9d1a5e4dc59cde4189ef5" gracePeriod=600 Oct 07 14:30:51 crc kubenswrapper[4959]: I1007 14:30:51.385372 4959 generic.go:334] "Generic (PLEG): container finished" podID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerID="f740f3b30981ad537e3b0957d254cc4251e26f8688f9d1a5e4dc59cde4189ef5" exitCode=0 Oct 07 14:30:51 crc kubenswrapper[4959]: I1007 14:30:51.385435 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerDied","Data":"f740f3b30981ad537e3b0957d254cc4251e26f8688f9d1a5e4dc59cde4189ef5"} Oct 07 14:30:51 crc kubenswrapper[4959]: I1007 14:30:51.385833 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerStarted","Data":"5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d"} Oct 07 14:30:51 crc kubenswrapper[4959]: I1007 14:30:51.385859 4959 scope.go:117] "RemoveContainer" containerID="0356775c31c3e2ae414b1122d13b3035bbcaddb886602ff9cacbb7ca8e029d7d" Oct 07 14:31:12 crc kubenswrapper[4959]: I1007 14:31:12.644766 4959 generic.go:334] "Generic (PLEG): container finished" podID="6645e33c-3472-41cd-8f39-a702210bc860" containerID="b9ed1f1b16adb671f43f9a2773821695ee0a65f46ef494ac485b899221601ef7" exitCode=0 Oct 07 14:31:12 crc kubenswrapper[4959]: I1007 14:31:12.645674 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" event={"ID":"6645e33c-3472-41cd-8f39-a702210bc860","Type":"ContainerDied","Data":"b9ed1f1b16adb671f43f9a2773821695ee0a65f46ef494ac485b899221601ef7"} Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.110046 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.162395 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlzbz\" (UniqueName: \"kubernetes.io/projected/6645e33c-3472-41cd-8f39-a702210bc860-kube-api-access-zlzbz\") pod \"6645e33c-3472-41cd-8f39-a702210bc860\" (UID: \"6645e33c-3472-41cd-8f39-a702210bc860\") " Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.162471 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-ssh-key\") pod \"6645e33c-3472-41cd-8f39-a702210bc860\" (UID: \"6645e33c-3472-41cd-8f39-a702210bc860\") " Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.162511 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-neutron-ovn-metadata-agent-neutron-config-0\") pod \"6645e33c-3472-41cd-8f39-a702210bc860\" (UID: \"6645e33c-3472-41cd-8f39-a702210bc860\") " Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.162569 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-nova-metadata-neutron-config-0\") pod \"6645e33c-3472-41cd-8f39-a702210bc860\" (UID: \"6645e33c-3472-41cd-8f39-a702210bc860\") " Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.162702 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-neutron-metadata-combined-ca-bundle\") pod \"6645e33c-3472-41cd-8f39-a702210bc860\" (UID: \"6645e33c-3472-41cd-8f39-a702210bc860\") " Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.162722 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-ceph\") pod \"6645e33c-3472-41cd-8f39-a702210bc860\" (UID: \"6645e33c-3472-41cd-8f39-a702210bc860\") " Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.162754 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-inventory\") pod \"6645e33c-3472-41cd-8f39-a702210bc860\" (UID: \"6645e33c-3472-41cd-8f39-a702210bc860\") " Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.171265 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-ceph" (OuterVolumeSpecName: "ceph") pod "6645e33c-3472-41cd-8f39-a702210bc860" (UID: "6645e33c-3472-41cd-8f39-a702210bc860"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.171756 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6645e33c-3472-41cd-8f39-a702210bc860-kube-api-access-zlzbz" (OuterVolumeSpecName: "kube-api-access-zlzbz") pod "6645e33c-3472-41cd-8f39-a702210bc860" (UID: "6645e33c-3472-41cd-8f39-a702210bc860"). InnerVolumeSpecName "kube-api-access-zlzbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.174611 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "6645e33c-3472-41cd-8f39-a702210bc860" (UID: "6645e33c-3472-41cd-8f39-a702210bc860"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.198599 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6645e33c-3472-41cd-8f39-a702210bc860" (UID: "6645e33c-3472-41cd-8f39-a702210bc860"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.199955 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-inventory" (OuterVolumeSpecName: "inventory") pod "6645e33c-3472-41cd-8f39-a702210bc860" (UID: "6645e33c-3472-41cd-8f39-a702210bc860"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.200653 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "6645e33c-3472-41cd-8f39-a702210bc860" (UID: "6645e33c-3472-41cd-8f39-a702210bc860"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.230254 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "6645e33c-3472-41cd-8f39-a702210bc860" (UID: "6645e33c-3472-41cd-8f39-a702210bc860"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.265992 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlzbz\" (UniqueName: \"kubernetes.io/projected/6645e33c-3472-41cd-8f39-a702210bc860-kube-api-access-zlzbz\") on node \"crc\" DevicePath \"\"" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.266393 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.266410 4959 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.266427 4959 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.266442 4959 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.266454 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-ceph\") on node \"crc\" DevicePath \"\"" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.266465 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6645e33c-3472-41cd-8f39-a702210bc860-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.673757 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.675439 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb" event={"ID":"6645e33c-3472-41cd-8f39-a702210bc860","Type":"ContainerDied","Data":"c523f365cdcb39573aacc4d986a6233aee1d7ad90df684a98e291b41914a9d60"} Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.675495 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c523f365cdcb39573aacc4d986a6233aee1d7ad90df684a98e291b41914a9d60" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.785736 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9"] Oct 07 14:31:14 crc kubenswrapper[4959]: E1007 14:31:14.786198 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6645e33c-3472-41cd-8f39-a702210bc860" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.786213 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="6645e33c-3472-41cd-8f39-a702210bc860" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.786402 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="6645e33c-3472-41cd-8f39-a702210bc860" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.787741 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.791944 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.792406 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.792495 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.792414 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.795800 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-c6vmh" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.796121 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.797882 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9"] Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.881913 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fswj9\" (UID: \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.882004 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fswj9\" (UID: \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.882142 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fswj9\" (UID: \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.882237 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fswj9\" (UID: \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.882295 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwjzg\" (UniqueName: \"kubernetes.io/projected/e5aa9a4d-4a36-481d-8853-c8a5b586d974-kube-api-access-kwjzg\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fswj9\" (UID: \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.882393 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fswj9\" (UID: \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.983636 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fswj9\" (UID: \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.983812 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fswj9\" (UID: \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.983922 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fswj9\" (UID: \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.983982 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwjzg\" (UniqueName: \"kubernetes.io/projected/e5aa9a4d-4a36-481d-8853-c8a5b586d974-kube-api-access-kwjzg\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fswj9\" (UID: \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.984064 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fswj9\" (UID: \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.984143 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fswj9\" (UID: \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.989958 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fswj9\" (UID: \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.990918 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fswj9\" (UID: \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.991048 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fswj9\" (UID: \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.991291 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fswj9\" (UID: \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9" Oct 07 14:31:14 crc kubenswrapper[4959]: I1007 14:31:14.991637 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fswj9\" (UID: \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9" Oct 07 14:31:15 crc kubenswrapper[4959]: I1007 14:31:15.009622 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwjzg\" (UniqueName: \"kubernetes.io/projected/e5aa9a4d-4a36-481d-8853-c8a5b586d974-kube-api-access-kwjzg\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fswj9\" (UID: \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9" Oct 07 14:31:15 crc kubenswrapper[4959]: I1007 14:31:15.125396 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9" Oct 07 14:31:15 crc kubenswrapper[4959]: I1007 14:31:15.706690 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9"] Oct 07 14:31:15 crc kubenswrapper[4959]: W1007 14:31:15.714821 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5aa9a4d_4a36_481d_8853_c8a5b586d974.slice/crio-d2a659c73ae60cccd0b5b254160f34130fb720fb165640011c243c36d9d7343d WatchSource:0}: Error finding container d2a659c73ae60cccd0b5b254160f34130fb720fb165640011c243c36d9d7343d: Status 404 returned error can't find the container with id d2a659c73ae60cccd0b5b254160f34130fb720fb165640011c243c36d9d7343d Oct 07 14:31:15 crc kubenswrapper[4959]: I1007 14:31:15.720497 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 14:31:16 crc kubenswrapper[4959]: I1007 14:31:16.693365 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9" event={"ID":"e5aa9a4d-4a36-481d-8853-c8a5b586d974","Type":"ContainerStarted","Data":"96c48e440493a3d15d8f96920621a17789215ad9c383b26ca0a0f7328d944d87"} Oct 07 14:31:16 crc kubenswrapper[4959]: I1007 14:31:16.693804 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9" event={"ID":"e5aa9a4d-4a36-481d-8853-c8a5b586d974","Type":"ContainerStarted","Data":"d2a659c73ae60cccd0b5b254160f34130fb720fb165640011c243c36d9d7343d"} Oct 07 14:31:16 crc kubenswrapper[4959]: I1007 14:31:16.712275 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9" podStartSLOduration=2.235453007 podStartE2EDuration="2.712251464s" podCreationTimestamp="2025-10-07 14:31:14 +0000 UTC" firstStartedPulling="2025-10-07 14:31:15.720195586 +0000 UTC m=+2757.803599901" lastFinishedPulling="2025-10-07 14:31:16.196994033 +0000 UTC m=+2758.280398358" observedRunningTime="2025-10-07 14:31:16.709982876 +0000 UTC m=+2758.793387201" watchObservedRunningTime="2025-10-07 14:31:16.712251464 +0000 UTC m=+2758.795655789" Oct 07 14:31:49 crc kubenswrapper[4959]: I1007 14:31:49.993013 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-827pq"] Oct 07 14:31:49 crc kubenswrapper[4959]: I1007 14:31:49.995995 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-827pq" Oct 07 14:31:50 crc kubenswrapper[4959]: I1007 14:31:50.019203 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-827pq"] Oct 07 14:31:50 crc kubenswrapper[4959]: I1007 14:31:50.107531 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b33d93b-ebf4-49b6-84be-ddd4d33b4c10-catalog-content\") pod \"certified-operators-827pq\" (UID: \"6b33d93b-ebf4-49b6-84be-ddd4d33b4c10\") " pod="openshift-marketplace/certified-operators-827pq" Oct 07 14:31:50 crc kubenswrapper[4959]: I1007 14:31:50.107887 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b33d93b-ebf4-49b6-84be-ddd4d33b4c10-utilities\") pod \"certified-operators-827pq\" (UID: \"6b33d93b-ebf4-49b6-84be-ddd4d33b4c10\") " pod="openshift-marketplace/certified-operators-827pq" Oct 07 14:31:50 crc kubenswrapper[4959]: I1007 14:31:50.107999 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b97t7\" (UniqueName: \"kubernetes.io/projected/6b33d93b-ebf4-49b6-84be-ddd4d33b4c10-kube-api-access-b97t7\") pod \"certified-operators-827pq\" (UID: \"6b33d93b-ebf4-49b6-84be-ddd4d33b4c10\") " pod="openshift-marketplace/certified-operators-827pq" Oct 07 14:31:50 crc kubenswrapper[4959]: I1007 14:31:50.209872 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b33d93b-ebf4-49b6-84be-ddd4d33b4c10-catalog-content\") pod \"certified-operators-827pq\" (UID: \"6b33d93b-ebf4-49b6-84be-ddd4d33b4c10\") " pod="openshift-marketplace/certified-operators-827pq" Oct 07 14:31:50 crc kubenswrapper[4959]: I1007 14:31:50.210006 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b33d93b-ebf4-49b6-84be-ddd4d33b4c10-utilities\") pod \"certified-operators-827pq\" (UID: \"6b33d93b-ebf4-49b6-84be-ddd4d33b4c10\") " pod="openshift-marketplace/certified-operators-827pq" Oct 07 14:31:50 crc kubenswrapper[4959]: I1007 14:31:50.210037 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b97t7\" (UniqueName: \"kubernetes.io/projected/6b33d93b-ebf4-49b6-84be-ddd4d33b4c10-kube-api-access-b97t7\") pod \"certified-operators-827pq\" (UID: \"6b33d93b-ebf4-49b6-84be-ddd4d33b4c10\") " pod="openshift-marketplace/certified-operators-827pq" Oct 07 14:31:50 crc kubenswrapper[4959]: I1007 14:31:50.210908 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b33d93b-ebf4-49b6-84be-ddd4d33b4c10-catalog-content\") pod \"certified-operators-827pq\" (UID: \"6b33d93b-ebf4-49b6-84be-ddd4d33b4c10\") " pod="openshift-marketplace/certified-operators-827pq" Oct 07 14:31:50 crc kubenswrapper[4959]: I1007 14:31:50.210929 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b33d93b-ebf4-49b6-84be-ddd4d33b4c10-utilities\") pod \"certified-operators-827pq\" (UID: \"6b33d93b-ebf4-49b6-84be-ddd4d33b4c10\") " pod="openshift-marketplace/certified-operators-827pq" Oct 07 14:31:50 crc kubenswrapper[4959]: I1007 14:31:50.231755 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b97t7\" (UniqueName: \"kubernetes.io/projected/6b33d93b-ebf4-49b6-84be-ddd4d33b4c10-kube-api-access-b97t7\") pod \"certified-operators-827pq\" (UID: \"6b33d93b-ebf4-49b6-84be-ddd4d33b4c10\") " pod="openshift-marketplace/certified-operators-827pq" Oct 07 14:31:50 crc kubenswrapper[4959]: I1007 14:31:50.339008 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-827pq" Oct 07 14:31:50 crc kubenswrapper[4959]: I1007 14:31:50.894865 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-827pq"] Oct 07 14:31:51 crc kubenswrapper[4959]: I1007 14:31:51.034535 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-827pq" event={"ID":"6b33d93b-ebf4-49b6-84be-ddd4d33b4c10","Type":"ContainerStarted","Data":"facfb671f4c31d00aa2b0e79f958f01fa1a664cd2afb9c26f34f39ae8ef71c61"} Oct 07 14:31:52 crc kubenswrapper[4959]: I1007 14:31:52.049399 4959 generic.go:334] "Generic (PLEG): container finished" podID="6b33d93b-ebf4-49b6-84be-ddd4d33b4c10" containerID="0542f7928d87afdb1c6139d7f0455649ef32b059cd3db7e1d1a087c839209a27" exitCode=0 Oct 07 14:31:52 crc kubenswrapper[4959]: I1007 14:31:52.049472 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-827pq" event={"ID":"6b33d93b-ebf4-49b6-84be-ddd4d33b4c10","Type":"ContainerDied","Data":"0542f7928d87afdb1c6139d7f0455649ef32b059cd3db7e1d1a087c839209a27"} Oct 07 14:31:54 crc kubenswrapper[4959]: I1007 14:31:54.070296 4959 generic.go:334] "Generic (PLEG): container finished" podID="6b33d93b-ebf4-49b6-84be-ddd4d33b4c10" containerID="8455d94a66a755d6cbf22c69067e8b17bc4d940b2159ad4b6d2b90928ad54baa" exitCode=0 Oct 07 14:31:54 crc kubenswrapper[4959]: I1007 14:31:54.070363 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-827pq" event={"ID":"6b33d93b-ebf4-49b6-84be-ddd4d33b4c10","Type":"ContainerDied","Data":"8455d94a66a755d6cbf22c69067e8b17bc4d940b2159ad4b6d2b90928ad54baa"} Oct 07 14:31:55 crc kubenswrapper[4959]: I1007 14:31:55.082279 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-827pq" event={"ID":"6b33d93b-ebf4-49b6-84be-ddd4d33b4c10","Type":"ContainerStarted","Data":"11506ac3d1fc5554ff66cbfaedaafba0e202df07e6f794c0ccee07c4a935621b"} Oct 07 14:31:55 crc kubenswrapper[4959]: I1007 14:31:55.108350 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-827pq" podStartSLOduration=3.5390994559999998 podStartE2EDuration="6.108324671s" podCreationTimestamp="2025-10-07 14:31:49 +0000 UTC" firstStartedPulling="2025-10-07 14:31:52.052760347 +0000 UTC m=+2794.136164702" lastFinishedPulling="2025-10-07 14:31:54.621985592 +0000 UTC m=+2796.705389917" observedRunningTime="2025-10-07 14:31:55.100291757 +0000 UTC m=+2797.183696092" watchObservedRunningTime="2025-10-07 14:31:55.108324671 +0000 UTC m=+2797.191728996" Oct 07 14:32:00 crc kubenswrapper[4959]: I1007 14:32:00.340173 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-827pq" Oct 07 14:32:00 crc kubenswrapper[4959]: I1007 14:32:00.341031 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-827pq" Oct 07 14:32:00 crc kubenswrapper[4959]: I1007 14:32:00.399948 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-827pq" Oct 07 14:32:01 crc kubenswrapper[4959]: I1007 14:32:01.203207 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-827pq" Oct 07 14:32:01 crc kubenswrapper[4959]: I1007 14:32:01.288173 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-827pq"] Oct 07 14:32:03 crc kubenswrapper[4959]: I1007 14:32:03.167935 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-827pq" podUID="6b33d93b-ebf4-49b6-84be-ddd4d33b4c10" containerName="registry-server" containerID="cri-o://11506ac3d1fc5554ff66cbfaedaafba0e202df07e6f794c0ccee07c4a935621b" gracePeriod=2 Oct 07 14:32:03 crc kubenswrapper[4959]: I1007 14:32:03.641254 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-827pq" Oct 07 14:32:03 crc kubenswrapper[4959]: I1007 14:32:03.743673 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b33d93b-ebf4-49b6-84be-ddd4d33b4c10-utilities\") pod \"6b33d93b-ebf4-49b6-84be-ddd4d33b4c10\" (UID: \"6b33d93b-ebf4-49b6-84be-ddd4d33b4c10\") " Oct 07 14:32:03 crc kubenswrapper[4959]: I1007 14:32:03.743768 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b97t7\" (UniqueName: \"kubernetes.io/projected/6b33d93b-ebf4-49b6-84be-ddd4d33b4c10-kube-api-access-b97t7\") pod \"6b33d93b-ebf4-49b6-84be-ddd4d33b4c10\" (UID: \"6b33d93b-ebf4-49b6-84be-ddd4d33b4c10\") " Oct 07 14:32:03 crc kubenswrapper[4959]: I1007 14:32:03.743938 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b33d93b-ebf4-49b6-84be-ddd4d33b4c10-catalog-content\") pod \"6b33d93b-ebf4-49b6-84be-ddd4d33b4c10\" (UID: \"6b33d93b-ebf4-49b6-84be-ddd4d33b4c10\") " Oct 07 14:32:03 crc kubenswrapper[4959]: I1007 14:32:03.746285 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b33d93b-ebf4-49b6-84be-ddd4d33b4c10-utilities" (OuterVolumeSpecName: "utilities") pod "6b33d93b-ebf4-49b6-84be-ddd4d33b4c10" (UID: "6b33d93b-ebf4-49b6-84be-ddd4d33b4c10"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:32:03 crc kubenswrapper[4959]: I1007 14:32:03.753529 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b33d93b-ebf4-49b6-84be-ddd4d33b4c10-kube-api-access-b97t7" (OuterVolumeSpecName: "kube-api-access-b97t7") pod "6b33d93b-ebf4-49b6-84be-ddd4d33b4c10" (UID: "6b33d93b-ebf4-49b6-84be-ddd4d33b4c10"). InnerVolumeSpecName "kube-api-access-b97t7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:32:03 crc kubenswrapper[4959]: I1007 14:32:03.847163 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b33d93b-ebf4-49b6-84be-ddd4d33b4c10-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 14:32:03 crc kubenswrapper[4959]: I1007 14:32:03.847272 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b97t7\" (UniqueName: \"kubernetes.io/projected/6b33d93b-ebf4-49b6-84be-ddd4d33b4c10-kube-api-access-b97t7\") on node \"crc\" DevicePath \"\"" Oct 07 14:32:04 crc kubenswrapper[4959]: I1007 14:32:04.182402 4959 generic.go:334] "Generic (PLEG): container finished" podID="6b33d93b-ebf4-49b6-84be-ddd4d33b4c10" containerID="11506ac3d1fc5554ff66cbfaedaafba0e202df07e6f794c0ccee07c4a935621b" exitCode=0 Oct 07 14:32:04 crc kubenswrapper[4959]: I1007 14:32:04.182473 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-827pq" event={"ID":"6b33d93b-ebf4-49b6-84be-ddd4d33b4c10","Type":"ContainerDied","Data":"11506ac3d1fc5554ff66cbfaedaafba0e202df07e6f794c0ccee07c4a935621b"} Oct 07 14:32:04 crc kubenswrapper[4959]: I1007 14:32:04.182527 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-827pq" Oct 07 14:32:04 crc kubenswrapper[4959]: I1007 14:32:04.182562 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-827pq" event={"ID":"6b33d93b-ebf4-49b6-84be-ddd4d33b4c10","Type":"ContainerDied","Data":"facfb671f4c31d00aa2b0e79f958f01fa1a664cd2afb9c26f34f39ae8ef71c61"} Oct 07 14:32:04 crc kubenswrapper[4959]: I1007 14:32:04.182596 4959 scope.go:117] "RemoveContainer" containerID="11506ac3d1fc5554ff66cbfaedaafba0e202df07e6f794c0ccee07c4a935621b" Oct 07 14:32:04 crc kubenswrapper[4959]: I1007 14:32:04.220515 4959 scope.go:117] "RemoveContainer" containerID="8455d94a66a755d6cbf22c69067e8b17bc4d940b2159ad4b6d2b90928ad54baa" Oct 07 14:32:04 crc kubenswrapper[4959]: I1007 14:32:04.255683 4959 scope.go:117] "RemoveContainer" containerID="0542f7928d87afdb1c6139d7f0455649ef32b059cd3db7e1d1a087c839209a27" Oct 07 14:32:04 crc kubenswrapper[4959]: I1007 14:32:04.305516 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b33d93b-ebf4-49b6-84be-ddd4d33b4c10-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6b33d93b-ebf4-49b6-84be-ddd4d33b4c10" (UID: "6b33d93b-ebf4-49b6-84be-ddd4d33b4c10"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:32:04 crc kubenswrapper[4959]: I1007 14:32:04.307231 4959 scope.go:117] "RemoveContainer" containerID="11506ac3d1fc5554ff66cbfaedaafba0e202df07e6f794c0ccee07c4a935621b" Oct 07 14:32:04 crc kubenswrapper[4959]: E1007 14:32:04.307963 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11506ac3d1fc5554ff66cbfaedaafba0e202df07e6f794c0ccee07c4a935621b\": container with ID starting with 11506ac3d1fc5554ff66cbfaedaafba0e202df07e6f794c0ccee07c4a935621b not found: ID does not exist" containerID="11506ac3d1fc5554ff66cbfaedaafba0e202df07e6f794c0ccee07c4a935621b" Oct 07 14:32:04 crc kubenswrapper[4959]: I1007 14:32:04.308016 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11506ac3d1fc5554ff66cbfaedaafba0e202df07e6f794c0ccee07c4a935621b"} err="failed to get container status \"11506ac3d1fc5554ff66cbfaedaafba0e202df07e6f794c0ccee07c4a935621b\": rpc error: code = NotFound desc = could not find container \"11506ac3d1fc5554ff66cbfaedaafba0e202df07e6f794c0ccee07c4a935621b\": container with ID starting with 11506ac3d1fc5554ff66cbfaedaafba0e202df07e6f794c0ccee07c4a935621b not found: ID does not exist" Oct 07 14:32:04 crc kubenswrapper[4959]: I1007 14:32:04.308053 4959 scope.go:117] "RemoveContainer" containerID="8455d94a66a755d6cbf22c69067e8b17bc4d940b2159ad4b6d2b90928ad54baa" Oct 07 14:32:04 crc kubenswrapper[4959]: E1007 14:32:04.308900 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8455d94a66a755d6cbf22c69067e8b17bc4d940b2159ad4b6d2b90928ad54baa\": container with ID starting with 8455d94a66a755d6cbf22c69067e8b17bc4d940b2159ad4b6d2b90928ad54baa not found: ID does not exist" containerID="8455d94a66a755d6cbf22c69067e8b17bc4d940b2159ad4b6d2b90928ad54baa" Oct 07 14:32:04 crc kubenswrapper[4959]: I1007 14:32:04.308963 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8455d94a66a755d6cbf22c69067e8b17bc4d940b2159ad4b6d2b90928ad54baa"} err="failed to get container status \"8455d94a66a755d6cbf22c69067e8b17bc4d940b2159ad4b6d2b90928ad54baa\": rpc error: code = NotFound desc = could not find container \"8455d94a66a755d6cbf22c69067e8b17bc4d940b2159ad4b6d2b90928ad54baa\": container with ID starting with 8455d94a66a755d6cbf22c69067e8b17bc4d940b2159ad4b6d2b90928ad54baa not found: ID does not exist" Oct 07 14:32:04 crc kubenswrapper[4959]: I1007 14:32:04.309011 4959 scope.go:117] "RemoveContainer" containerID="0542f7928d87afdb1c6139d7f0455649ef32b059cd3db7e1d1a087c839209a27" Oct 07 14:32:04 crc kubenswrapper[4959]: E1007 14:32:04.309603 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0542f7928d87afdb1c6139d7f0455649ef32b059cd3db7e1d1a087c839209a27\": container with ID starting with 0542f7928d87afdb1c6139d7f0455649ef32b059cd3db7e1d1a087c839209a27 not found: ID does not exist" containerID="0542f7928d87afdb1c6139d7f0455649ef32b059cd3db7e1d1a087c839209a27" Oct 07 14:32:04 crc kubenswrapper[4959]: I1007 14:32:04.309633 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0542f7928d87afdb1c6139d7f0455649ef32b059cd3db7e1d1a087c839209a27"} err="failed to get container status \"0542f7928d87afdb1c6139d7f0455649ef32b059cd3db7e1d1a087c839209a27\": rpc error: code = NotFound desc = could not find container \"0542f7928d87afdb1c6139d7f0455649ef32b059cd3db7e1d1a087c839209a27\": container with ID starting with 0542f7928d87afdb1c6139d7f0455649ef32b059cd3db7e1d1a087c839209a27 not found: ID does not exist" Oct 07 14:32:04 crc kubenswrapper[4959]: I1007 14:32:04.358804 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b33d93b-ebf4-49b6-84be-ddd4d33b4c10-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 14:32:04 crc kubenswrapper[4959]: I1007 14:32:04.547289 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-827pq"] Oct 07 14:32:04 crc kubenswrapper[4959]: I1007 14:32:04.560448 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-827pq"] Oct 07 14:32:04 crc kubenswrapper[4959]: I1007 14:32:04.670383 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b33d93b-ebf4-49b6-84be-ddd4d33b4c10" path="/var/lib/kubelet/pods/6b33d93b-ebf4-49b6-84be-ddd4d33b4c10/volumes" Oct 07 14:32:47 crc kubenswrapper[4959]: I1007 14:32:47.187825 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-n228c"] Oct 07 14:32:47 crc kubenswrapper[4959]: E1007 14:32:47.189145 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b33d93b-ebf4-49b6-84be-ddd4d33b4c10" containerName="extract-content" Oct 07 14:32:47 crc kubenswrapper[4959]: I1007 14:32:47.189163 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b33d93b-ebf4-49b6-84be-ddd4d33b4c10" containerName="extract-content" Oct 07 14:32:47 crc kubenswrapper[4959]: E1007 14:32:47.189187 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b33d93b-ebf4-49b6-84be-ddd4d33b4c10" containerName="extract-utilities" Oct 07 14:32:47 crc kubenswrapper[4959]: I1007 14:32:47.189195 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b33d93b-ebf4-49b6-84be-ddd4d33b4c10" containerName="extract-utilities" Oct 07 14:32:47 crc kubenswrapper[4959]: E1007 14:32:47.189232 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b33d93b-ebf4-49b6-84be-ddd4d33b4c10" containerName="registry-server" Oct 07 14:32:47 crc kubenswrapper[4959]: I1007 14:32:47.189239 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b33d93b-ebf4-49b6-84be-ddd4d33b4c10" containerName="registry-server" Oct 07 14:32:47 crc kubenswrapper[4959]: I1007 14:32:47.189418 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b33d93b-ebf4-49b6-84be-ddd4d33b4c10" containerName="registry-server" Oct 07 14:32:47 crc kubenswrapper[4959]: I1007 14:32:47.190808 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n228c" Oct 07 14:32:47 crc kubenswrapper[4959]: I1007 14:32:47.207695 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n228c"] Oct 07 14:32:47 crc kubenswrapper[4959]: I1007 14:32:47.317548 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5565d858-1a96-4bb7-9645-0d47c41e5246-catalog-content\") pod \"redhat-operators-n228c\" (UID: \"5565d858-1a96-4bb7-9645-0d47c41e5246\") " pod="openshift-marketplace/redhat-operators-n228c" Oct 07 14:32:47 crc kubenswrapper[4959]: I1007 14:32:47.317718 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrt8b\" (UniqueName: \"kubernetes.io/projected/5565d858-1a96-4bb7-9645-0d47c41e5246-kube-api-access-vrt8b\") pod \"redhat-operators-n228c\" (UID: \"5565d858-1a96-4bb7-9645-0d47c41e5246\") " pod="openshift-marketplace/redhat-operators-n228c" Oct 07 14:32:47 crc kubenswrapper[4959]: I1007 14:32:47.317754 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5565d858-1a96-4bb7-9645-0d47c41e5246-utilities\") pod \"redhat-operators-n228c\" (UID: \"5565d858-1a96-4bb7-9645-0d47c41e5246\") " pod="openshift-marketplace/redhat-operators-n228c" Oct 07 14:32:47 crc kubenswrapper[4959]: I1007 14:32:47.420352 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5565d858-1a96-4bb7-9645-0d47c41e5246-catalog-content\") pod \"redhat-operators-n228c\" (UID: \"5565d858-1a96-4bb7-9645-0d47c41e5246\") " pod="openshift-marketplace/redhat-operators-n228c" Oct 07 14:32:47 crc kubenswrapper[4959]: I1007 14:32:47.420496 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrt8b\" (UniqueName: \"kubernetes.io/projected/5565d858-1a96-4bb7-9645-0d47c41e5246-kube-api-access-vrt8b\") pod \"redhat-operators-n228c\" (UID: \"5565d858-1a96-4bb7-9645-0d47c41e5246\") " pod="openshift-marketplace/redhat-operators-n228c" Oct 07 14:32:47 crc kubenswrapper[4959]: I1007 14:32:47.420524 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5565d858-1a96-4bb7-9645-0d47c41e5246-utilities\") pod \"redhat-operators-n228c\" (UID: \"5565d858-1a96-4bb7-9645-0d47c41e5246\") " pod="openshift-marketplace/redhat-operators-n228c" Oct 07 14:32:47 crc kubenswrapper[4959]: I1007 14:32:47.421052 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5565d858-1a96-4bb7-9645-0d47c41e5246-catalog-content\") pod \"redhat-operators-n228c\" (UID: \"5565d858-1a96-4bb7-9645-0d47c41e5246\") " pod="openshift-marketplace/redhat-operators-n228c" Oct 07 14:32:47 crc kubenswrapper[4959]: I1007 14:32:47.421176 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5565d858-1a96-4bb7-9645-0d47c41e5246-utilities\") pod \"redhat-operators-n228c\" (UID: \"5565d858-1a96-4bb7-9645-0d47c41e5246\") " pod="openshift-marketplace/redhat-operators-n228c" Oct 07 14:32:47 crc kubenswrapper[4959]: I1007 14:32:47.444908 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrt8b\" (UniqueName: \"kubernetes.io/projected/5565d858-1a96-4bb7-9645-0d47c41e5246-kube-api-access-vrt8b\") pod \"redhat-operators-n228c\" (UID: \"5565d858-1a96-4bb7-9645-0d47c41e5246\") " pod="openshift-marketplace/redhat-operators-n228c" Oct 07 14:32:47 crc kubenswrapper[4959]: I1007 14:32:47.529587 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n228c" Oct 07 14:32:48 crc kubenswrapper[4959]: I1007 14:32:48.042724 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n228c"] Oct 07 14:32:48 crc kubenswrapper[4959]: I1007 14:32:48.654995 4959 generic.go:334] "Generic (PLEG): container finished" podID="5565d858-1a96-4bb7-9645-0d47c41e5246" containerID="ba7f938cc841091909d89641bbf58c08cc198c76dacac1ba4b22a8d4bf5a2e16" exitCode=0 Oct 07 14:32:48 crc kubenswrapper[4959]: I1007 14:32:48.668688 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n228c" event={"ID":"5565d858-1a96-4bb7-9645-0d47c41e5246","Type":"ContainerDied","Data":"ba7f938cc841091909d89641bbf58c08cc198c76dacac1ba4b22a8d4bf5a2e16"} Oct 07 14:32:48 crc kubenswrapper[4959]: I1007 14:32:48.668744 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n228c" event={"ID":"5565d858-1a96-4bb7-9645-0d47c41e5246","Type":"ContainerStarted","Data":"b4f4c0fd2589d67c3a486e242491bce121f51c9dfae04a6576d7faf85bb6ae55"} Oct 07 14:32:49 crc kubenswrapper[4959]: I1007 14:32:49.586276 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-l9v7g"] Oct 07 14:32:49 crc kubenswrapper[4959]: I1007 14:32:49.590593 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l9v7g" Oct 07 14:32:49 crc kubenswrapper[4959]: I1007 14:32:49.601197 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l9v7g"] Oct 07 14:32:49 crc kubenswrapper[4959]: I1007 14:32:49.688231 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1e12add-9b0e-481c-888d-d4c4a19bdf30-utilities\") pod \"redhat-marketplace-l9v7g\" (UID: \"d1e12add-9b0e-481c-888d-d4c4a19bdf30\") " pod="openshift-marketplace/redhat-marketplace-l9v7g" Oct 07 14:32:49 crc kubenswrapper[4959]: I1007 14:32:49.688574 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwcjf\" (UniqueName: \"kubernetes.io/projected/d1e12add-9b0e-481c-888d-d4c4a19bdf30-kube-api-access-lwcjf\") pod \"redhat-marketplace-l9v7g\" (UID: \"d1e12add-9b0e-481c-888d-d4c4a19bdf30\") " pod="openshift-marketplace/redhat-marketplace-l9v7g" Oct 07 14:32:49 crc kubenswrapper[4959]: I1007 14:32:49.688677 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1e12add-9b0e-481c-888d-d4c4a19bdf30-catalog-content\") pod \"redhat-marketplace-l9v7g\" (UID: \"d1e12add-9b0e-481c-888d-d4c4a19bdf30\") " pod="openshift-marketplace/redhat-marketplace-l9v7g" Oct 07 14:32:49 crc kubenswrapper[4959]: I1007 14:32:49.790844 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1e12add-9b0e-481c-888d-d4c4a19bdf30-utilities\") pod \"redhat-marketplace-l9v7g\" (UID: \"d1e12add-9b0e-481c-888d-d4c4a19bdf30\") " pod="openshift-marketplace/redhat-marketplace-l9v7g" Oct 07 14:32:49 crc kubenswrapper[4959]: I1007 14:32:49.791082 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwcjf\" (UniqueName: \"kubernetes.io/projected/d1e12add-9b0e-481c-888d-d4c4a19bdf30-kube-api-access-lwcjf\") pod \"redhat-marketplace-l9v7g\" (UID: \"d1e12add-9b0e-481c-888d-d4c4a19bdf30\") " pod="openshift-marketplace/redhat-marketplace-l9v7g" Oct 07 14:32:49 crc kubenswrapper[4959]: I1007 14:32:49.791164 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1e12add-9b0e-481c-888d-d4c4a19bdf30-catalog-content\") pod \"redhat-marketplace-l9v7g\" (UID: \"d1e12add-9b0e-481c-888d-d4c4a19bdf30\") " pod="openshift-marketplace/redhat-marketplace-l9v7g" Oct 07 14:32:49 crc kubenswrapper[4959]: I1007 14:32:49.791408 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1e12add-9b0e-481c-888d-d4c4a19bdf30-utilities\") pod \"redhat-marketplace-l9v7g\" (UID: \"d1e12add-9b0e-481c-888d-d4c4a19bdf30\") " pod="openshift-marketplace/redhat-marketplace-l9v7g" Oct 07 14:32:49 crc kubenswrapper[4959]: I1007 14:32:49.791988 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1e12add-9b0e-481c-888d-d4c4a19bdf30-catalog-content\") pod \"redhat-marketplace-l9v7g\" (UID: \"d1e12add-9b0e-481c-888d-d4c4a19bdf30\") " pod="openshift-marketplace/redhat-marketplace-l9v7g" Oct 07 14:32:49 crc kubenswrapper[4959]: I1007 14:32:49.813159 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwcjf\" (UniqueName: \"kubernetes.io/projected/d1e12add-9b0e-481c-888d-d4c4a19bdf30-kube-api-access-lwcjf\") pod \"redhat-marketplace-l9v7g\" (UID: \"d1e12add-9b0e-481c-888d-d4c4a19bdf30\") " pod="openshift-marketplace/redhat-marketplace-l9v7g" Oct 07 14:32:49 crc kubenswrapper[4959]: I1007 14:32:49.953424 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l9v7g" Oct 07 14:32:50 crc kubenswrapper[4959]: W1007 14:32:50.299148 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1e12add_9b0e_481c_888d_d4c4a19bdf30.slice/crio-489b2d929815d76c318e21cc59c1ffca4e518d771d4f60e58737d092050d3c24 WatchSource:0}: Error finding container 489b2d929815d76c318e21cc59c1ffca4e518d771d4f60e58737d092050d3c24: Status 404 returned error can't find the container with id 489b2d929815d76c318e21cc59c1ffca4e518d771d4f60e58737d092050d3c24 Oct 07 14:32:50 crc kubenswrapper[4959]: I1007 14:32:50.301383 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l9v7g"] Oct 07 14:32:50 crc kubenswrapper[4959]: I1007 14:32:50.630281 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:32:50 crc kubenswrapper[4959]: I1007 14:32:50.630353 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:32:50 crc kubenswrapper[4959]: I1007 14:32:50.688213 4959 generic.go:334] "Generic (PLEG): container finished" podID="d1e12add-9b0e-481c-888d-d4c4a19bdf30" containerID="7cf016f02a5454ad010aa9b2f03801c00b9a04c4f385aa3223394ffa1cd1f283" exitCode=0 Oct 07 14:32:50 crc kubenswrapper[4959]: I1007 14:32:50.688356 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l9v7g" event={"ID":"d1e12add-9b0e-481c-888d-d4c4a19bdf30","Type":"ContainerDied","Data":"7cf016f02a5454ad010aa9b2f03801c00b9a04c4f385aa3223394ffa1cd1f283"} Oct 07 14:32:50 crc kubenswrapper[4959]: I1007 14:32:50.688407 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l9v7g" event={"ID":"d1e12add-9b0e-481c-888d-d4c4a19bdf30","Type":"ContainerStarted","Data":"489b2d929815d76c318e21cc59c1ffca4e518d771d4f60e58737d092050d3c24"} Oct 07 14:32:50 crc kubenswrapper[4959]: I1007 14:32:50.691248 4959 generic.go:334] "Generic (PLEG): container finished" podID="5565d858-1a96-4bb7-9645-0d47c41e5246" containerID="7b84646c6eee470acdae6b8ac7a0870629aa90ae8d128036034a8981a272301c" exitCode=0 Oct 07 14:32:50 crc kubenswrapper[4959]: I1007 14:32:50.691280 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n228c" event={"ID":"5565d858-1a96-4bb7-9645-0d47c41e5246","Type":"ContainerDied","Data":"7b84646c6eee470acdae6b8ac7a0870629aa90ae8d128036034a8981a272301c"} Oct 07 14:32:51 crc kubenswrapper[4959]: I1007 14:32:51.704234 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n228c" event={"ID":"5565d858-1a96-4bb7-9645-0d47c41e5246","Type":"ContainerStarted","Data":"70745b1740c6cf567d6917a5b99cd2dc88e7c538825b394e4ead9bace8d46f7e"} Oct 07 14:32:51 crc kubenswrapper[4959]: I1007 14:32:51.707832 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l9v7g" event={"ID":"d1e12add-9b0e-481c-888d-d4c4a19bdf30","Type":"ContainerStarted","Data":"41176a418e8fefd82a2a2122d0350fdc267c7260cb9811172a34e4a71e8f2a3f"} Oct 07 14:32:51 crc kubenswrapper[4959]: I1007 14:32:51.733238 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-n228c" podStartSLOduration=2.102614495 podStartE2EDuration="4.733215992s" podCreationTimestamp="2025-10-07 14:32:47 +0000 UTC" firstStartedPulling="2025-10-07 14:32:48.662908976 +0000 UTC m=+2850.746313311" lastFinishedPulling="2025-10-07 14:32:51.293510483 +0000 UTC m=+2853.376914808" observedRunningTime="2025-10-07 14:32:51.730358099 +0000 UTC m=+2853.813762444" watchObservedRunningTime="2025-10-07 14:32:51.733215992 +0000 UTC m=+2853.816620317" Oct 07 14:32:52 crc kubenswrapper[4959]: I1007 14:32:52.719006 4959 generic.go:334] "Generic (PLEG): container finished" podID="d1e12add-9b0e-481c-888d-d4c4a19bdf30" containerID="41176a418e8fefd82a2a2122d0350fdc267c7260cb9811172a34e4a71e8f2a3f" exitCode=0 Oct 07 14:32:52 crc kubenswrapper[4959]: I1007 14:32:52.719070 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l9v7g" event={"ID":"d1e12add-9b0e-481c-888d-d4c4a19bdf30","Type":"ContainerDied","Data":"41176a418e8fefd82a2a2122d0350fdc267c7260cb9811172a34e4a71e8f2a3f"} Oct 07 14:32:53 crc kubenswrapper[4959]: I1007 14:32:53.731515 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l9v7g" event={"ID":"d1e12add-9b0e-481c-888d-d4c4a19bdf30","Type":"ContainerStarted","Data":"f68bcc907414fcd9618b020b062267f5f9923fb9442936ee2dc8b1d2b210aa18"} Oct 07 14:32:53 crc kubenswrapper[4959]: I1007 14:32:53.758603 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-l9v7g" podStartSLOduration=2.3071522780000002 podStartE2EDuration="4.75854123s" podCreationTimestamp="2025-10-07 14:32:49 +0000 UTC" firstStartedPulling="2025-10-07 14:32:50.690299947 +0000 UTC m=+2852.773704272" lastFinishedPulling="2025-10-07 14:32:53.141688899 +0000 UTC m=+2855.225093224" observedRunningTime="2025-10-07 14:32:53.753381529 +0000 UTC m=+2855.836785864" watchObservedRunningTime="2025-10-07 14:32:53.75854123 +0000 UTC m=+2855.841945565" Oct 07 14:32:57 crc kubenswrapper[4959]: I1007 14:32:57.530071 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-n228c" Oct 07 14:32:57 crc kubenswrapper[4959]: I1007 14:32:57.530966 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-n228c" Oct 07 14:32:57 crc kubenswrapper[4959]: I1007 14:32:57.590365 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-n228c" Oct 07 14:32:57 crc kubenswrapper[4959]: I1007 14:32:57.835906 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-n228c" Oct 07 14:32:58 crc kubenswrapper[4959]: I1007 14:32:58.967235 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n228c"] Oct 07 14:32:59 crc kubenswrapper[4959]: I1007 14:32:59.788865 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-n228c" podUID="5565d858-1a96-4bb7-9645-0d47c41e5246" containerName="registry-server" containerID="cri-o://70745b1740c6cf567d6917a5b99cd2dc88e7c538825b394e4ead9bace8d46f7e" gracePeriod=2 Oct 07 14:32:59 crc kubenswrapper[4959]: I1007 14:32:59.954300 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-l9v7g" Oct 07 14:32:59 crc kubenswrapper[4959]: I1007 14:32:59.954613 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-l9v7g" Oct 07 14:33:00 crc kubenswrapper[4959]: I1007 14:33:00.041421 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-l9v7g" Oct 07 14:33:00 crc kubenswrapper[4959]: I1007 14:33:00.268456 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n228c" Oct 07 14:33:00 crc kubenswrapper[4959]: I1007 14:33:00.325395 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5565d858-1a96-4bb7-9645-0d47c41e5246-utilities\") pod \"5565d858-1a96-4bb7-9645-0d47c41e5246\" (UID: \"5565d858-1a96-4bb7-9645-0d47c41e5246\") " Oct 07 14:33:00 crc kubenswrapper[4959]: I1007 14:33:00.325523 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5565d858-1a96-4bb7-9645-0d47c41e5246-catalog-content\") pod \"5565d858-1a96-4bb7-9645-0d47c41e5246\" (UID: \"5565d858-1a96-4bb7-9645-0d47c41e5246\") " Oct 07 14:33:00 crc kubenswrapper[4959]: I1007 14:33:00.325736 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrt8b\" (UniqueName: \"kubernetes.io/projected/5565d858-1a96-4bb7-9645-0d47c41e5246-kube-api-access-vrt8b\") pod \"5565d858-1a96-4bb7-9645-0d47c41e5246\" (UID: \"5565d858-1a96-4bb7-9645-0d47c41e5246\") " Oct 07 14:33:00 crc kubenswrapper[4959]: I1007 14:33:00.329310 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5565d858-1a96-4bb7-9645-0d47c41e5246-utilities" (OuterVolumeSpecName: "utilities") pod "5565d858-1a96-4bb7-9645-0d47c41e5246" (UID: "5565d858-1a96-4bb7-9645-0d47c41e5246"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:33:00 crc kubenswrapper[4959]: I1007 14:33:00.334439 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5565d858-1a96-4bb7-9645-0d47c41e5246-kube-api-access-vrt8b" (OuterVolumeSpecName: "kube-api-access-vrt8b") pod "5565d858-1a96-4bb7-9645-0d47c41e5246" (UID: "5565d858-1a96-4bb7-9645-0d47c41e5246"). InnerVolumeSpecName "kube-api-access-vrt8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:33:00 crc kubenswrapper[4959]: I1007 14:33:00.429781 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5565d858-1a96-4bb7-9645-0d47c41e5246-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 14:33:00 crc kubenswrapper[4959]: I1007 14:33:00.430150 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrt8b\" (UniqueName: \"kubernetes.io/projected/5565d858-1a96-4bb7-9645-0d47c41e5246-kube-api-access-vrt8b\") on node \"crc\" DevicePath \"\"" Oct 07 14:33:00 crc kubenswrapper[4959]: I1007 14:33:00.456440 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5565d858-1a96-4bb7-9645-0d47c41e5246-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5565d858-1a96-4bb7-9645-0d47c41e5246" (UID: "5565d858-1a96-4bb7-9645-0d47c41e5246"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:33:00 crc kubenswrapper[4959]: I1007 14:33:00.531828 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5565d858-1a96-4bb7-9645-0d47c41e5246-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 14:33:00 crc kubenswrapper[4959]: I1007 14:33:00.812199 4959 generic.go:334] "Generic (PLEG): container finished" podID="5565d858-1a96-4bb7-9645-0d47c41e5246" containerID="70745b1740c6cf567d6917a5b99cd2dc88e7c538825b394e4ead9bace8d46f7e" exitCode=0 Oct 07 14:33:00 crc kubenswrapper[4959]: I1007 14:33:00.813279 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n228c" Oct 07 14:33:00 crc kubenswrapper[4959]: I1007 14:33:00.813753 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n228c" event={"ID":"5565d858-1a96-4bb7-9645-0d47c41e5246","Type":"ContainerDied","Data":"70745b1740c6cf567d6917a5b99cd2dc88e7c538825b394e4ead9bace8d46f7e"} Oct 07 14:33:00 crc kubenswrapper[4959]: I1007 14:33:00.813783 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n228c" event={"ID":"5565d858-1a96-4bb7-9645-0d47c41e5246","Type":"ContainerDied","Data":"b4f4c0fd2589d67c3a486e242491bce121f51c9dfae04a6576d7faf85bb6ae55"} Oct 07 14:33:00 crc kubenswrapper[4959]: I1007 14:33:00.813800 4959 scope.go:117] "RemoveContainer" containerID="70745b1740c6cf567d6917a5b99cd2dc88e7c538825b394e4ead9bace8d46f7e" Oct 07 14:33:00 crc kubenswrapper[4959]: I1007 14:33:00.864287 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n228c"] Oct 07 14:33:00 crc kubenswrapper[4959]: I1007 14:33:00.867106 4959 scope.go:117] "RemoveContainer" containerID="7b84646c6eee470acdae6b8ac7a0870629aa90ae8d128036034a8981a272301c" Oct 07 14:33:00 crc kubenswrapper[4959]: I1007 14:33:00.878406 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-n228c"] Oct 07 14:33:00 crc kubenswrapper[4959]: I1007 14:33:00.893630 4959 scope.go:117] "RemoveContainer" containerID="ba7f938cc841091909d89641bbf58c08cc198c76dacac1ba4b22a8d4bf5a2e16" Oct 07 14:33:00 crc kubenswrapper[4959]: I1007 14:33:00.900785 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-l9v7g" Oct 07 14:33:00 crc kubenswrapper[4959]: I1007 14:33:00.951862 4959 scope.go:117] "RemoveContainer" containerID="70745b1740c6cf567d6917a5b99cd2dc88e7c538825b394e4ead9bace8d46f7e" Oct 07 14:33:00 crc kubenswrapper[4959]: E1007 14:33:00.952415 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70745b1740c6cf567d6917a5b99cd2dc88e7c538825b394e4ead9bace8d46f7e\": container with ID starting with 70745b1740c6cf567d6917a5b99cd2dc88e7c538825b394e4ead9bace8d46f7e not found: ID does not exist" containerID="70745b1740c6cf567d6917a5b99cd2dc88e7c538825b394e4ead9bace8d46f7e" Oct 07 14:33:00 crc kubenswrapper[4959]: I1007 14:33:00.952485 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70745b1740c6cf567d6917a5b99cd2dc88e7c538825b394e4ead9bace8d46f7e"} err="failed to get container status \"70745b1740c6cf567d6917a5b99cd2dc88e7c538825b394e4ead9bace8d46f7e\": rpc error: code = NotFound desc = could not find container \"70745b1740c6cf567d6917a5b99cd2dc88e7c538825b394e4ead9bace8d46f7e\": container with ID starting with 70745b1740c6cf567d6917a5b99cd2dc88e7c538825b394e4ead9bace8d46f7e not found: ID does not exist" Oct 07 14:33:00 crc kubenswrapper[4959]: I1007 14:33:00.952536 4959 scope.go:117] "RemoveContainer" containerID="7b84646c6eee470acdae6b8ac7a0870629aa90ae8d128036034a8981a272301c" Oct 07 14:33:00 crc kubenswrapper[4959]: E1007 14:33:00.953678 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b84646c6eee470acdae6b8ac7a0870629aa90ae8d128036034a8981a272301c\": container with ID starting with 7b84646c6eee470acdae6b8ac7a0870629aa90ae8d128036034a8981a272301c not found: ID does not exist" containerID="7b84646c6eee470acdae6b8ac7a0870629aa90ae8d128036034a8981a272301c" Oct 07 14:33:00 crc kubenswrapper[4959]: I1007 14:33:00.953874 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b84646c6eee470acdae6b8ac7a0870629aa90ae8d128036034a8981a272301c"} err="failed to get container status \"7b84646c6eee470acdae6b8ac7a0870629aa90ae8d128036034a8981a272301c\": rpc error: code = NotFound desc = could not find container \"7b84646c6eee470acdae6b8ac7a0870629aa90ae8d128036034a8981a272301c\": container with ID starting with 7b84646c6eee470acdae6b8ac7a0870629aa90ae8d128036034a8981a272301c not found: ID does not exist" Oct 07 14:33:00 crc kubenswrapper[4959]: I1007 14:33:00.954036 4959 scope.go:117] "RemoveContainer" containerID="ba7f938cc841091909d89641bbf58c08cc198c76dacac1ba4b22a8d4bf5a2e16" Oct 07 14:33:00 crc kubenswrapper[4959]: E1007 14:33:00.954762 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba7f938cc841091909d89641bbf58c08cc198c76dacac1ba4b22a8d4bf5a2e16\": container with ID starting with ba7f938cc841091909d89641bbf58c08cc198c76dacac1ba4b22a8d4bf5a2e16 not found: ID does not exist" containerID="ba7f938cc841091909d89641bbf58c08cc198c76dacac1ba4b22a8d4bf5a2e16" Oct 07 14:33:00 crc kubenswrapper[4959]: I1007 14:33:00.954833 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba7f938cc841091909d89641bbf58c08cc198c76dacac1ba4b22a8d4bf5a2e16"} err="failed to get container status \"ba7f938cc841091909d89641bbf58c08cc198c76dacac1ba4b22a8d4bf5a2e16\": rpc error: code = NotFound desc = could not find container \"ba7f938cc841091909d89641bbf58c08cc198c76dacac1ba4b22a8d4bf5a2e16\": container with ID starting with ba7f938cc841091909d89641bbf58c08cc198c76dacac1ba4b22a8d4bf5a2e16 not found: ID does not exist" Oct 07 14:33:02 crc kubenswrapper[4959]: I1007 14:33:02.674035 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5565d858-1a96-4bb7-9645-0d47c41e5246" path="/var/lib/kubelet/pods/5565d858-1a96-4bb7-9645-0d47c41e5246/volumes" Oct 07 14:33:02 crc kubenswrapper[4959]: I1007 14:33:02.968167 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l9v7g"] Oct 07 14:33:02 crc kubenswrapper[4959]: I1007 14:33:02.968406 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-l9v7g" podUID="d1e12add-9b0e-481c-888d-d4c4a19bdf30" containerName="registry-server" containerID="cri-o://f68bcc907414fcd9618b020b062267f5f9923fb9442936ee2dc8b1d2b210aa18" gracePeriod=2 Oct 07 14:33:03 crc kubenswrapper[4959]: I1007 14:33:03.504750 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l9v7g" Oct 07 14:33:03 crc kubenswrapper[4959]: I1007 14:33:03.597496 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwcjf\" (UniqueName: \"kubernetes.io/projected/d1e12add-9b0e-481c-888d-d4c4a19bdf30-kube-api-access-lwcjf\") pod \"d1e12add-9b0e-481c-888d-d4c4a19bdf30\" (UID: \"d1e12add-9b0e-481c-888d-d4c4a19bdf30\") " Oct 07 14:33:03 crc kubenswrapper[4959]: I1007 14:33:03.597663 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1e12add-9b0e-481c-888d-d4c4a19bdf30-catalog-content\") pod \"d1e12add-9b0e-481c-888d-d4c4a19bdf30\" (UID: \"d1e12add-9b0e-481c-888d-d4c4a19bdf30\") " Oct 07 14:33:03 crc kubenswrapper[4959]: I1007 14:33:03.597702 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1e12add-9b0e-481c-888d-d4c4a19bdf30-utilities\") pod \"d1e12add-9b0e-481c-888d-d4c4a19bdf30\" (UID: \"d1e12add-9b0e-481c-888d-d4c4a19bdf30\") " Oct 07 14:33:03 crc kubenswrapper[4959]: I1007 14:33:03.599920 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1e12add-9b0e-481c-888d-d4c4a19bdf30-utilities" (OuterVolumeSpecName: "utilities") pod "d1e12add-9b0e-481c-888d-d4c4a19bdf30" (UID: "d1e12add-9b0e-481c-888d-d4c4a19bdf30"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:33:03 crc kubenswrapper[4959]: I1007 14:33:03.611118 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1e12add-9b0e-481c-888d-d4c4a19bdf30-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d1e12add-9b0e-481c-888d-d4c4a19bdf30" (UID: "d1e12add-9b0e-481c-888d-d4c4a19bdf30"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:33:03 crc kubenswrapper[4959]: I1007 14:33:03.624441 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1e12add-9b0e-481c-888d-d4c4a19bdf30-kube-api-access-lwcjf" (OuterVolumeSpecName: "kube-api-access-lwcjf") pod "d1e12add-9b0e-481c-888d-d4c4a19bdf30" (UID: "d1e12add-9b0e-481c-888d-d4c4a19bdf30"). InnerVolumeSpecName "kube-api-access-lwcjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:33:03 crc kubenswrapper[4959]: I1007 14:33:03.699625 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwcjf\" (UniqueName: \"kubernetes.io/projected/d1e12add-9b0e-481c-888d-d4c4a19bdf30-kube-api-access-lwcjf\") on node \"crc\" DevicePath \"\"" Oct 07 14:33:03 crc kubenswrapper[4959]: I1007 14:33:03.699659 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1e12add-9b0e-481c-888d-d4c4a19bdf30-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 14:33:03 crc kubenswrapper[4959]: I1007 14:33:03.699668 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1e12add-9b0e-481c-888d-d4c4a19bdf30-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 14:33:03 crc kubenswrapper[4959]: I1007 14:33:03.847542 4959 generic.go:334] "Generic (PLEG): container finished" podID="d1e12add-9b0e-481c-888d-d4c4a19bdf30" containerID="f68bcc907414fcd9618b020b062267f5f9923fb9442936ee2dc8b1d2b210aa18" exitCode=0 Oct 07 14:33:03 crc kubenswrapper[4959]: I1007 14:33:03.847649 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l9v7g" event={"ID":"d1e12add-9b0e-481c-888d-d4c4a19bdf30","Type":"ContainerDied","Data":"f68bcc907414fcd9618b020b062267f5f9923fb9442936ee2dc8b1d2b210aa18"} Oct 07 14:33:03 crc kubenswrapper[4959]: I1007 14:33:03.848008 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l9v7g" event={"ID":"d1e12add-9b0e-481c-888d-d4c4a19bdf30","Type":"ContainerDied","Data":"489b2d929815d76c318e21cc59c1ffca4e518d771d4f60e58737d092050d3c24"} Oct 07 14:33:03 crc kubenswrapper[4959]: I1007 14:33:03.848072 4959 scope.go:117] "RemoveContainer" containerID="f68bcc907414fcd9618b020b062267f5f9923fb9442936ee2dc8b1d2b210aa18" Oct 07 14:33:03 crc kubenswrapper[4959]: I1007 14:33:03.847670 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l9v7g" Oct 07 14:33:03 crc kubenswrapper[4959]: I1007 14:33:03.873306 4959 scope.go:117] "RemoveContainer" containerID="41176a418e8fefd82a2a2122d0350fdc267c7260cb9811172a34e4a71e8f2a3f" Oct 07 14:33:03 crc kubenswrapper[4959]: I1007 14:33:03.890982 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l9v7g"] Oct 07 14:33:03 crc kubenswrapper[4959]: I1007 14:33:03.899470 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-l9v7g"] Oct 07 14:33:03 crc kubenswrapper[4959]: I1007 14:33:03.922014 4959 scope.go:117] "RemoveContainer" containerID="7cf016f02a5454ad010aa9b2f03801c00b9a04c4f385aa3223394ffa1cd1f283" Oct 07 14:33:03 crc kubenswrapper[4959]: I1007 14:33:03.967542 4959 scope.go:117] "RemoveContainer" containerID="f68bcc907414fcd9618b020b062267f5f9923fb9442936ee2dc8b1d2b210aa18" Oct 07 14:33:03 crc kubenswrapper[4959]: E1007 14:33:03.969525 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f68bcc907414fcd9618b020b062267f5f9923fb9442936ee2dc8b1d2b210aa18\": container with ID starting with f68bcc907414fcd9618b020b062267f5f9923fb9442936ee2dc8b1d2b210aa18 not found: ID does not exist" containerID="f68bcc907414fcd9618b020b062267f5f9923fb9442936ee2dc8b1d2b210aa18" Oct 07 14:33:03 crc kubenswrapper[4959]: I1007 14:33:03.969598 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f68bcc907414fcd9618b020b062267f5f9923fb9442936ee2dc8b1d2b210aa18"} err="failed to get container status \"f68bcc907414fcd9618b020b062267f5f9923fb9442936ee2dc8b1d2b210aa18\": rpc error: code = NotFound desc = could not find container \"f68bcc907414fcd9618b020b062267f5f9923fb9442936ee2dc8b1d2b210aa18\": container with ID starting with f68bcc907414fcd9618b020b062267f5f9923fb9442936ee2dc8b1d2b210aa18 not found: ID does not exist" Oct 07 14:33:03 crc kubenswrapper[4959]: I1007 14:33:03.969638 4959 scope.go:117] "RemoveContainer" containerID="41176a418e8fefd82a2a2122d0350fdc267c7260cb9811172a34e4a71e8f2a3f" Oct 07 14:33:03 crc kubenswrapper[4959]: E1007 14:33:03.969948 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41176a418e8fefd82a2a2122d0350fdc267c7260cb9811172a34e4a71e8f2a3f\": container with ID starting with 41176a418e8fefd82a2a2122d0350fdc267c7260cb9811172a34e4a71e8f2a3f not found: ID does not exist" containerID="41176a418e8fefd82a2a2122d0350fdc267c7260cb9811172a34e4a71e8f2a3f" Oct 07 14:33:03 crc kubenswrapper[4959]: I1007 14:33:03.969979 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41176a418e8fefd82a2a2122d0350fdc267c7260cb9811172a34e4a71e8f2a3f"} err="failed to get container status \"41176a418e8fefd82a2a2122d0350fdc267c7260cb9811172a34e4a71e8f2a3f\": rpc error: code = NotFound desc = could not find container \"41176a418e8fefd82a2a2122d0350fdc267c7260cb9811172a34e4a71e8f2a3f\": container with ID starting with 41176a418e8fefd82a2a2122d0350fdc267c7260cb9811172a34e4a71e8f2a3f not found: ID does not exist" Oct 07 14:33:03 crc kubenswrapper[4959]: I1007 14:33:03.969995 4959 scope.go:117] "RemoveContainer" containerID="7cf016f02a5454ad010aa9b2f03801c00b9a04c4f385aa3223394ffa1cd1f283" Oct 07 14:33:03 crc kubenswrapper[4959]: E1007 14:33:03.970343 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cf016f02a5454ad010aa9b2f03801c00b9a04c4f385aa3223394ffa1cd1f283\": container with ID starting with 7cf016f02a5454ad010aa9b2f03801c00b9a04c4f385aa3223394ffa1cd1f283 not found: ID does not exist" containerID="7cf016f02a5454ad010aa9b2f03801c00b9a04c4f385aa3223394ffa1cd1f283" Oct 07 14:33:03 crc kubenswrapper[4959]: I1007 14:33:03.970379 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cf016f02a5454ad010aa9b2f03801c00b9a04c4f385aa3223394ffa1cd1f283"} err="failed to get container status \"7cf016f02a5454ad010aa9b2f03801c00b9a04c4f385aa3223394ffa1cd1f283\": rpc error: code = NotFound desc = could not find container \"7cf016f02a5454ad010aa9b2f03801c00b9a04c4f385aa3223394ffa1cd1f283\": container with ID starting with 7cf016f02a5454ad010aa9b2f03801c00b9a04c4f385aa3223394ffa1cd1f283 not found: ID does not exist" Oct 07 14:33:04 crc kubenswrapper[4959]: I1007 14:33:04.666450 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1e12add-9b0e-481c-888d-d4c4a19bdf30" path="/var/lib/kubelet/pods/d1e12add-9b0e-481c-888d-d4c4a19bdf30/volumes" Oct 07 14:33:20 crc kubenswrapper[4959]: I1007 14:33:20.630729 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:33:20 crc kubenswrapper[4959]: I1007 14:33:20.631515 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:33:50 crc kubenswrapper[4959]: I1007 14:33:50.629666 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:33:50 crc kubenswrapper[4959]: I1007 14:33:50.630452 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:33:50 crc kubenswrapper[4959]: I1007 14:33:50.630511 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 14:33:50 crc kubenswrapper[4959]: I1007 14:33:50.631521 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d"} pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 14:33:50 crc kubenswrapper[4959]: I1007 14:33:50.631591 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" containerID="cri-o://5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d" gracePeriod=600 Oct 07 14:33:50 crc kubenswrapper[4959]: E1007 14:33:50.815507 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:33:51 crc kubenswrapper[4959]: I1007 14:33:51.274866 4959 generic.go:334] "Generic (PLEG): container finished" podID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerID="5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d" exitCode=0 Oct 07 14:33:51 crc kubenswrapper[4959]: I1007 14:33:51.274936 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerDied","Data":"5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d"} Oct 07 14:33:51 crc kubenswrapper[4959]: I1007 14:33:51.275332 4959 scope.go:117] "RemoveContainer" containerID="f740f3b30981ad537e3b0957d254cc4251e26f8688f9d1a5e4dc59cde4189ef5" Oct 07 14:33:51 crc kubenswrapper[4959]: I1007 14:33:51.275883 4959 scope.go:117] "RemoveContainer" containerID="5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d" Oct 07 14:33:51 crc kubenswrapper[4959]: E1007 14:33:51.276287 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:34:05 crc kubenswrapper[4959]: I1007 14:34:05.654175 4959 scope.go:117] "RemoveContainer" containerID="5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d" Oct 07 14:34:05 crc kubenswrapper[4959]: E1007 14:34:05.655285 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:34:17 crc kubenswrapper[4959]: I1007 14:34:17.654033 4959 scope.go:117] "RemoveContainer" containerID="5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d" Oct 07 14:34:17 crc kubenswrapper[4959]: E1007 14:34:17.655090 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:34:30 crc kubenswrapper[4959]: I1007 14:34:30.654014 4959 scope.go:117] "RemoveContainer" containerID="5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d" Oct 07 14:34:30 crc kubenswrapper[4959]: E1007 14:34:30.655165 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:34:31 crc kubenswrapper[4959]: I1007 14:34:31.942512 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5zxkd"] Oct 07 14:34:31 crc kubenswrapper[4959]: E1007 14:34:31.943400 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1e12add-9b0e-481c-888d-d4c4a19bdf30" containerName="extract-content" Oct 07 14:34:31 crc kubenswrapper[4959]: I1007 14:34:31.943419 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1e12add-9b0e-481c-888d-d4c4a19bdf30" containerName="extract-content" Oct 07 14:34:31 crc kubenswrapper[4959]: E1007 14:34:31.943439 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5565d858-1a96-4bb7-9645-0d47c41e5246" containerName="extract-utilities" Oct 07 14:34:31 crc kubenswrapper[4959]: I1007 14:34:31.943448 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5565d858-1a96-4bb7-9645-0d47c41e5246" containerName="extract-utilities" Oct 07 14:34:31 crc kubenswrapper[4959]: E1007 14:34:31.943465 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5565d858-1a96-4bb7-9645-0d47c41e5246" containerName="registry-server" Oct 07 14:34:31 crc kubenswrapper[4959]: I1007 14:34:31.943473 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5565d858-1a96-4bb7-9645-0d47c41e5246" containerName="registry-server" Oct 07 14:34:31 crc kubenswrapper[4959]: E1007 14:34:31.943491 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5565d858-1a96-4bb7-9645-0d47c41e5246" containerName="extract-content" Oct 07 14:34:31 crc kubenswrapper[4959]: I1007 14:34:31.943497 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5565d858-1a96-4bb7-9645-0d47c41e5246" containerName="extract-content" Oct 07 14:34:31 crc kubenswrapper[4959]: E1007 14:34:31.943509 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1e12add-9b0e-481c-888d-d4c4a19bdf30" containerName="extract-utilities" Oct 07 14:34:31 crc kubenswrapper[4959]: I1007 14:34:31.943515 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1e12add-9b0e-481c-888d-d4c4a19bdf30" containerName="extract-utilities" Oct 07 14:34:31 crc kubenswrapper[4959]: E1007 14:34:31.943526 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1e12add-9b0e-481c-888d-d4c4a19bdf30" containerName="registry-server" Oct 07 14:34:31 crc kubenswrapper[4959]: I1007 14:34:31.943531 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1e12add-9b0e-481c-888d-d4c4a19bdf30" containerName="registry-server" Oct 07 14:34:31 crc kubenswrapper[4959]: I1007 14:34:31.943692 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="5565d858-1a96-4bb7-9645-0d47c41e5246" containerName="registry-server" Oct 07 14:34:31 crc kubenswrapper[4959]: I1007 14:34:31.943717 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1e12add-9b0e-481c-888d-d4c4a19bdf30" containerName="registry-server" Oct 07 14:34:31 crc kubenswrapper[4959]: I1007 14:34:31.945073 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5zxkd" Oct 07 14:34:31 crc kubenswrapper[4959]: I1007 14:34:31.961205 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5zxkd"] Oct 07 14:34:32 crc kubenswrapper[4959]: I1007 14:34:32.114030 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8c18b36-24da-43b2-8fc1-95a45bef266b-utilities\") pod \"community-operators-5zxkd\" (UID: \"b8c18b36-24da-43b2-8fc1-95a45bef266b\") " pod="openshift-marketplace/community-operators-5zxkd" Oct 07 14:34:32 crc kubenswrapper[4959]: I1007 14:34:32.114454 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8c18b36-24da-43b2-8fc1-95a45bef266b-catalog-content\") pod \"community-operators-5zxkd\" (UID: \"b8c18b36-24da-43b2-8fc1-95a45bef266b\") " pod="openshift-marketplace/community-operators-5zxkd" Oct 07 14:34:32 crc kubenswrapper[4959]: I1007 14:34:32.114489 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lvsq\" (UniqueName: \"kubernetes.io/projected/b8c18b36-24da-43b2-8fc1-95a45bef266b-kube-api-access-7lvsq\") pod \"community-operators-5zxkd\" (UID: \"b8c18b36-24da-43b2-8fc1-95a45bef266b\") " pod="openshift-marketplace/community-operators-5zxkd" Oct 07 14:34:32 crc kubenswrapper[4959]: I1007 14:34:32.216011 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8c18b36-24da-43b2-8fc1-95a45bef266b-utilities\") pod \"community-operators-5zxkd\" (UID: \"b8c18b36-24da-43b2-8fc1-95a45bef266b\") " pod="openshift-marketplace/community-operators-5zxkd" Oct 07 14:34:32 crc kubenswrapper[4959]: I1007 14:34:32.216074 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8c18b36-24da-43b2-8fc1-95a45bef266b-catalog-content\") pod \"community-operators-5zxkd\" (UID: \"b8c18b36-24da-43b2-8fc1-95a45bef266b\") " pod="openshift-marketplace/community-operators-5zxkd" Oct 07 14:34:32 crc kubenswrapper[4959]: I1007 14:34:32.216122 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lvsq\" (UniqueName: \"kubernetes.io/projected/b8c18b36-24da-43b2-8fc1-95a45bef266b-kube-api-access-7lvsq\") pod \"community-operators-5zxkd\" (UID: \"b8c18b36-24da-43b2-8fc1-95a45bef266b\") " pod="openshift-marketplace/community-operators-5zxkd" Oct 07 14:34:32 crc kubenswrapper[4959]: I1007 14:34:32.217289 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8c18b36-24da-43b2-8fc1-95a45bef266b-utilities\") pod \"community-operators-5zxkd\" (UID: \"b8c18b36-24da-43b2-8fc1-95a45bef266b\") " pod="openshift-marketplace/community-operators-5zxkd" Oct 07 14:34:32 crc kubenswrapper[4959]: I1007 14:34:32.217673 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8c18b36-24da-43b2-8fc1-95a45bef266b-catalog-content\") pod \"community-operators-5zxkd\" (UID: \"b8c18b36-24da-43b2-8fc1-95a45bef266b\") " pod="openshift-marketplace/community-operators-5zxkd" Oct 07 14:34:32 crc kubenswrapper[4959]: I1007 14:34:32.243374 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lvsq\" (UniqueName: \"kubernetes.io/projected/b8c18b36-24da-43b2-8fc1-95a45bef266b-kube-api-access-7lvsq\") pod \"community-operators-5zxkd\" (UID: \"b8c18b36-24da-43b2-8fc1-95a45bef266b\") " pod="openshift-marketplace/community-operators-5zxkd" Oct 07 14:34:32 crc kubenswrapper[4959]: I1007 14:34:32.266532 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5zxkd" Oct 07 14:34:32 crc kubenswrapper[4959]: I1007 14:34:32.948851 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5zxkd"] Oct 07 14:34:33 crc kubenswrapper[4959]: I1007 14:34:33.689256 4959 generic.go:334] "Generic (PLEG): container finished" podID="b8c18b36-24da-43b2-8fc1-95a45bef266b" containerID="d571e62225f2fda0432b8116152f8618ca1c699afd076e4d5d7cff8b4008b22d" exitCode=0 Oct 07 14:34:33 crc kubenswrapper[4959]: I1007 14:34:33.689448 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5zxkd" event={"ID":"b8c18b36-24da-43b2-8fc1-95a45bef266b","Type":"ContainerDied","Data":"d571e62225f2fda0432b8116152f8618ca1c699afd076e4d5d7cff8b4008b22d"} Oct 07 14:34:33 crc kubenswrapper[4959]: I1007 14:34:33.689576 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5zxkd" event={"ID":"b8c18b36-24da-43b2-8fc1-95a45bef266b","Type":"ContainerStarted","Data":"9d8159d2f093ac18abffed88c94d2911a1b8314a0072e2fb3d20d044ddc9ba3a"} Oct 07 14:34:35 crc kubenswrapper[4959]: I1007 14:34:35.708087 4959 generic.go:334] "Generic (PLEG): container finished" podID="b8c18b36-24da-43b2-8fc1-95a45bef266b" containerID="3a818ed075738a4576dfe07a2f1ddddfa9b33f724d51a0fc5a6b2b7aee9c26d6" exitCode=0 Oct 07 14:34:35 crc kubenswrapper[4959]: I1007 14:34:35.708145 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5zxkd" event={"ID":"b8c18b36-24da-43b2-8fc1-95a45bef266b","Type":"ContainerDied","Data":"3a818ed075738a4576dfe07a2f1ddddfa9b33f724d51a0fc5a6b2b7aee9c26d6"} Oct 07 14:34:36 crc kubenswrapper[4959]: I1007 14:34:36.719507 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5zxkd" event={"ID":"b8c18b36-24da-43b2-8fc1-95a45bef266b","Type":"ContainerStarted","Data":"aa60dd0cacd6822e0833da17690dd95c800d68f11a2ddb8ce5b415c256d94d3c"} Oct 07 14:34:36 crc kubenswrapper[4959]: I1007 14:34:36.744023 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5zxkd" podStartSLOduration=3.285348021 podStartE2EDuration="5.744001557s" podCreationTimestamp="2025-10-07 14:34:31 +0000 UTC" firstStartedPulling="2025-10-07 14:34:33.691786458 +0000 UTC m=+2955.775190783" lastFinishedPulling="2025-10-07 14:34:36.150440004 +0000 UTC m=+2958.233844319" observedRunningTime="2025-10-07 14:34:36.739018741 +0000 UTC m=+2958.822423066" watchObservedRunningTime="2025-10-07 14:34:36.744001557 +0000 UTC m=+2958.827405882" Oct 07 14:34:42 crc kubenswrapper[4959]: I1007 14:34:42.268008 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5zxkd" Oct 07 14:34:42 crc kubenswrapper[4959]: I1007 14:34:42.268514 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5zxkd" Oct 07 14:34:42 crc kubenswrapper[4959]: I1007 14:34:42.323328 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5zxkd" Oct 07 14:34:42 crc kubenswrapper[4959]: I1007 14:34:42.653429 4959 scope.go:117] "RemoveContainer" containerID="5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d" Oct 07 14:34:42 crc kubenswrapper[4959]: E1007 14:34:42.653709 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:34:42 crc kubenswrapper[4959]: I1007 14:34:42.818214 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5zxkd" Oct 07 14:34:42 crc kubenswrapper[4959]: I1007 14:34:42.866987 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5zxkd"] Oct 07 14:34:44 crc kubenswrapper[4959]: I1007 14:34:44.784762 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5zxkd" podUID="b8c18b36-24da-43b2-8fc1-95a45bef266b" containerName="registry-server" containerID="cri-o://aa60dd0cacd6822e0833da17690dd95c800d68f11a2ddb8ce5b415c256d94d3c" gracePeriod=2 Oct 07 14:34:45 crc kubenswrapper[4959]: I1007 14:34:45.291849 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5zxkd" Oct 07 14:34:45 crc kubenswrapper[4959]: I1007 14:34:45.405629 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lvsq\" (UniqueName: \"kubernetes.io/projected/b8c18b36-24da-43b2-8fc1-95a45bef266b-kube-api-access-7lvsq\") pod \"b8c18b36-24da-43b2-8fc1-95a45bef266b\" (UID: \"b8c18b36-24da-43b2-8fc1-95a45bef266b\") " Oct 07 14:34:45 crc kubenswrapper[4959]: I1007 14:34:45.406273 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8c18b36-24da-43b2-8fc1-95a45bef266b-utilities\") pod \"b8c18b36-24da-43b2-8fc1-95a45bef266b\" (UID: \"b8c18b36-24da-43b2-8fc1-95a45bef266b\") " Oct 07 14:34:45 crc kubenswrapper[4959]: I1007 14:34:45.406303 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8c18b36-24da-43b2-8fc1-95a45bef266b-catalog-content\") pod \"b8c18b36-24da-43b2-8fc1-95a45bef266b\" (UID: \"b8c18b36-24da-43b2-8fc1-95a45bef266b\") " Oct 07 14:34:45 crc kubenswrapper[4959]: I1007 14:34:45.407257 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8c18b36-24da-43b2-8fc1-95a45bef266b-utilities" (OuterVolumeSpecName: "utilities") pod "b8c18b36-24da-43b2-8fc1-95a45bef266b" (UID: "b8c18b36-24da-43b2-8fc1-95a45bef266b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:34:45 crc kubenswrapper[4959]: I1007 14:34:45.413691 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8c18b36-24da-43b2-8fc1-95a45bef266b-kube-api-access-7lvsq" (OuterVolumeSpecName: "kube-api-access-7lvsq") pod "b8c18b36-24da-43b2-8fc1-95a45bef266b" (UID: "b8c18b36-24da-43b2-8fc1-95a45bef266b"). InnerVolumeSpecName "kube-api-access-7lvsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:34:45 crc kubenswrapper[4959]: I1007 14:34:45.468467 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8c18b36-24da-43b2-8fc1-95a45bef266b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b8c18b36-24da-43b2-8fc1-95a45bef266b" (UID: "b8c18b36-24da-43b2-8fc1-95a45bef266b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:34:45 crc kubenswrapper[4959]: I1007 14:34:45.508899 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8c18b36-24da-43b2-8fc1-95a45bef266b-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 14:34:45 crc kubenswrapper[4959]: I1007 14:34:45.509396 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8c18b36-24da-43b2-8fc1-95a45bef266b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 14:34:45 crc kubenswrapper[4959]: I1007 14:34:45.509521 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lvsq\" (UniqueName: \"kubernetes.io/projected/b8c18b36-24da-43b2-8fc1-95a45bef266b-kube-api-access-7lvsq\") on node \"crc\" DevicePath \"\"" Oct 07 14:34:45 crc kubenswrapper[4959]: I1007 14:34:45.799277 4959 generic.go:334] "Generic (PLEG): container finished" podID="b8c18b36-24da-43b2-8fc1-95a45bef266b" containerID="aa60dd0cacd6822e0833da17690dd95c800d68f11a2ddb8ce5b415c256d94d3c" exitCode=0 Oct 07 14:34:45 crc kubenswrapper[4959]: I1007 14:34:45.799329 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5zxkd" event={"ID":"b8c18b36-24da-43b2-8fc1-95a45bef266b","Type":"ContainerDied","Data":"aa60dd0cacd6822e0833da17690dd95c800d68f11a2ddb8ce5b415c256d94d3c"} Oct 07 14:34:45 crc kubenswrapper[4959]: I1007 14:34:45.800238 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5zxkd" event={"ID":"b8c18b36-24da-43b2-8fc1-95a45bef266b","Type":"ContainerDied","Data":"9d8159d2f093ac18abffed88c94d2911a1b8314a0072e2fb3d20d044ddc9ba3a"} Oct 07 14:34:45 crc kubenswrapper[4959]: I1007 14:34:45.799387 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5zxkd" Oct 07 14:34:45 crc kubenswrapper[4959]: I1007 14:34:45.800270 4959 scope.go:117] "RemoveContainer" containerID="aa60dd0cacd6822e0833da17690dd95c800d68f11a2ddb8ce5b415c256d94d3c" Oct 07 14:34:45 crc kubenswrapper[4959]: I1007 14:34:45.830501 4959 scope.go:117] "RemoveContainer" containerID="3a818ed075738a4576dfe07a2f1ddddfa9b33f724d51a0fc5a6b2b7aee9c26d6" Oct 07 14:34:45 crc kubenswrapper[4959]: I1007 14:34:45.837278 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5zxkd"] Oct 07 14:34:45 crc kubenswrapper[4959]: I1007 14:34:45.848253 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5zxkd"] Oct 07 14:34:45 crc kubenswrapper[4959]: I1007 14:34:45.871912 4959 scope.go:117] "RemoveContainer" containerID="d571e62225f2fda0432b8116152f8618ca1c699afd076e4d5d7cff8b4008b22d" Oct 07 14:34:45 crc kubenswrapper[4959]: I1007 14:34:45.912721 4959 scope.go:117] "RemoveContainer" containerID="aa60dd0cacd6822e0833da17690dd95c800d68f11a2ddb8ce5b415c256d94d3c" Oct 07 14:34:45 crc kubenswrapper[4959]: E1007 14:34:45.914627 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa60dd0cacd6822e0833da17690dd95c800d68f11a2ddb8ce5b415c256d94d3c\": container with ID starting with aa60dd0cacd6822e0833da17690dd95c800d68f11a2ddb8ce5b415c256d94d3c not found: ID does not exist" containerID="aa60dd0cacd6822e0833da17690dd95c800d68f11a2ddb8ce5b415c256d94d3c" Oct 07 14:34:45 crc kubenswrapper[4959]: I1007 14:34:45.914734 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa60dd0cacd6822e0833da17690dd95c800d68f11a2ddb8ce5b415c256d94d3c"} err="failed to get container status \"aa60dd0cacd6822e0833da17690dd95c800d68f11a2ddb8ce5b415c256d94d3c\": rpc error: code = NotFound desc = could not find container \"aa60dd0cacd6822e0833da17690dd95c800d68f11a2ddb8ce5b415c256d94d3c\": container with ID starting with aa60dd0cacd6822e0833da17690dd95c800d68f11a2ddb8ce5b415c256d94d3c not found: ID does not exist" Oct 07 14:34:45 crc kubenswrapper[4959]: I1007 14:34:45.914801 4959 scope.go:117] "RemoveContainer" containerID="3a818ed075738a4576dfe07a2f1ddddfa9b33f724d51a0fc5a6b2b7aee9c26d6" Oct 07 14:34:45 crc kubenswrapper[4959]: E1007 14:34:45.915387 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a818ed075738a4576dfe07a2f1ddddfa9b33f724d51a0fc5a6b2b7aee9c26d6\": container with ID starting with 3a818ed075738a4576dfe07a2f1ddddfa9b33f724d51a0fc5a6b2b7aee9c26d6 not found: ID does not exist" containerID="3a818ed075738a4576dfe07a2f1ddddfa9b33f724d51a0fc5a6b2b7aee9c26d6" Oct 07 14:34:45 crc kubenswrapper[4959]: I1007 14:34:45.915417 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a818ed075738a4576dfe07a2f1ddddfa9b33f724d51a0fc5a6b2b7aee9c26d6"} err="failed to get container status \"3a818ed075738a4576dfe07a2f1ddddfa9b33f724d51a0fc5a6b2b7aee9c26d6\": rpc error: code = NotFound desc = could not find container \"3a818ed075738a4576dfe07a2f1ddddfa9b33f724d51a0fc5a6b2b7aee9c26d6\": container with ID starting with 3a818ed075738a4576dfe07a2f1ddddfa9b33f724d51a0fc5a6b2b7aee9c26d6 not found: ID does not exist" Oct 07 14:34:45 crc kubenswrapper[4959]: I1007 14:34:45.915441 4959 scope.go:117] "RemoveContainer" containerID="d571e62225f2fda0432b8116152f8618ca1c699afd076e4d5d7cff8b4008b22d" Oct 07 14:34:45 crc kubenswrapper[4959]: E1007 14:34:45.916092 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d571e62225f2fda0432b8116152f8618ca1c699afd076e4d5d7cff8b4008b22d\": container with ID starting with d571e62225f2fda0432b8116152f8618ca1c699afd076e4d5d7cff8b4008b22d not found: ID does not exist" containerID="d571e62225f2fda0432b8116152f8618ca1c699afd076e4d5d7cff8b4008b22d" Oct 07 14:34:45 crc kubenswrapper[4959]: I1007 14:34:45.916181 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d571e62225f2fda0432b8116152f8618ca1c699afd076e4d5d7cff8b4008b22d"} err="failed to get container status \"d571e62225f2fda0432b8116152f8618ca1c699afd076e4d5d7cff8b4008b22d\": rpc error: code = NotFound desc = could not find container \"d571e62225f2fda0432b8116152f8618ca1c699afd076e4d5d7cff8b4008b22d\": container with ID starting with d571e62225f2fda0432b8116152f8618ca1c699afd076e4d5d7cff8b4008b22d not found: ID does not exist" Oct 07 14:34:46 crc kubenswrapper[4959]: I1007 14:34:46.665488 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8c18b36-24da-43b2-8fc1-95a45bef266b" path="/var/lib/kubelet/pods/b8c18b36-24da-43b2-8fc1-95a45bef266b/volumes" Oct 07 14:34:56 crc kubenswrapper[4959]: I1007 14:34:56.654510 4959 scope.go:117] "RemoveContainer" containerID="5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d" Oct 07 14:34:56 crc kubenswrapper[4959]: E1007 14:34:56.655278 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:35:10 crc kubenswrapper[4959]: I1007 14:35:10.654395 4959 scope.go:117] "RemoveContainer" containerID="5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d" Oct 07 14:35:10 crc kubenswrapper[4959]: E1007 14:35:10.655435 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:35:24 crc kubenswrapper[4959]: I1007 14:35:24.654454 4959 scope.go:117] "RemoveContainer" containerID="5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d" Oct 07 14:35:24 crc kubenswrapper[4959]: E1007 14:35:24.655439 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:35:35 crc kubenswrapper[4959]: I1007 14:35:35.654043 4959 scope.go:117] "RemoveContainer" containerID="5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d" Oct 07 14:35:35 crc kubenswrapper[4959]: E1007 14:35:35.655787 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:35:43 crc kubenswrapper[4959]: I1007 14:35:43.390037 4959 generic.go:334] "Generic (PLEG): container finished" podID="e5aa9a4d-4a36-481d-8853-c8a5b586d974" containerID="96c48e440493a3d15d8f96920621a17789215ad9c383b26ca0a0f7328d944d87" exitCode=0 Oct 07 14:35:43 crc kubenswrapper[4959]: I1007 14:35:43.390224 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9" event={"ID":"e5aa9a4d-4a36-481d-8853-c8a5b586d974","Type":"ContainerDied","Data":"96c48e440493a3d15d8f96920621a17789215ad9c383b26ca0a0f7328d944d87"} Oct 07 14:35:44 crc kubenswrapper[4959]: I1007 14:35:44.835293 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9" Oct 07 14:35:44 crc kubenswrapper[4959]: I1007 14:35:44.914720 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwjzg\" (UniqueName: \"kubernetes.io/projected/e5aa9a4d-4a36-481d-8853-c8a5b586d974-kube-api-access-kwjzg\") pod \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\" (UID: \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\") " Oct 07 14:35:44 crc kubenswrapper[4959]: I1007 14:35:44.914890 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-libvirt-combined-ca-bundle\") pod \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\" (UID: \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\") " Oct 07 14:35:44 crc kubenswrapper[4959]: I1007 14:35:44.914969 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-ceph\") pod \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\" (UID: \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\") " Oct 07 14:35:44 crc kubenswrapper[4959]: I1007 14:35:44.915009 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-inventory\") pod \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\" (UID: \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\") " Oct 07 14:35:44 crc kubenswrapper[4959]: I1007 14:35:44.915157 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-libvirt-secret-0\") pod \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\" (UID: \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\") " Oct 07 14:35:44 crc kubenswrapper[4959]: I1007 14:35:44.915223 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-ssh-key\") pod \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\" (UID: \"e5aa9a4d-4a36-481d-8853-c8a5b586d974\") " Oct 07 14:35:44 crc kubenswrapper[4959]: I1007 14:35:44.922675 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5aa9a4d-4a36-481d-8853-c8a5b586d974-kube-api-access-kwjzg" (OuterVolumeSpecName: "kube-api-access-kwjzg") pod "e5aa9a4d-4a36-481d-8853-c8a5b586d974" (UID: "e5aa9a4d-4a36-481d-8853-c8a5b586d974"). InnerVolumeSpecName "kube-api-access-kwjzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:35:44 crc kubenswrapper[4959]: I1007 14:35:44.929868 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "e5aa9a4d-4a36-481d-8853-c8a5b586d974" (UID: "e5aa9a4d-4a36-481d-8853-c8a5b586d974"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:35:44 crc kubenswrapper[4959]: I1007 14:35:44.931573 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-ceph" (OuterVolumeSpecName: "ceph") pod "e5aa9a4d-4a36-481d-8853-c8a5b586d974" (UID: "e5aa9a4d-4a36-481d-8853-c8a5b586d974"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:35:44 crc kubenswrapper[4959]: I1007 14:35:44.949296 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e5aa9a4d-4a36-481d-8853-c8a5b586d974" (UID: "e5aa9a4d-4a36-481d-8853-c8a5b586d974"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:35:44 crc kubenswrapper[4959]: I1007 14:35:44.952028 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "e5aa9a4d-4a36-481d-8853-c8a5b586d974" (UID: "e5aa9a4d-4a36-481d-8853-c8a5b586d974"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:35:44 crc kubenswrapper[4959]: I1007 14:35:44.955939 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-inventory" (OuterVolumeSpecName: "inventory") pod "e5aa9a4d-4a36-481d-8853-c8a5b586d974" (UID: "e5aa9a4d-4a36-481d-8853-c8a5b586d974"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.017968 4959 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.018009 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-ceph\") on node \"crc\" DevicePath \"\"" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.018022 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.019580 4959 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.019609 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5aa9a4d-4a36-481d-8853-c8a5b586d974-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.019621 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwjzg\" (UniqueName: \"kubernetes.io/projected/e5aa9a4d-4a36-481d-8853-c8a5b586d974-kube-api-access-kwjzg\") on node \"crc\" DevicePath \"\"" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.410910 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9" event={"ID":"e5aa9a4d-4a36-481d-8853-c8a5b586d974","Type":"ContainerDied","Data":"d2a659c73ae60cccd0b5b254160f34130fb720fb165640011c243c36d9d7343d"} Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.410965 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2a659c73ae60cccd0b5b254160f34130fb720fb165640011c243c36d9d7343d" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.410969 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fswj9" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.526766 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh"] Oct 07 14:35:45 crc kubenswrapper[4959]: E1007 14:35:45.527543 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8c18b36-24da-43b2-8fc1-95a45bef266b" containerName="extract-utilities" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.527560 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8c18b36-24da-43b2-8fc1-95a45bef266b" containerName="extract-utilities" Oct 07 14:35:45 crc kubenswrapper[4959]: E1007 14:35:45.527602 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8c18b36-24da-43b2-8fc1-95a45bef266b" containerName="registry-server" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.527610 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8c18b36-24da-43b2-8fc1-95a45bef266b" containerName="registry-server" Oct 07 14:35:45 crc kubenswrapper[4959]: E1007 14:35:45.527622 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8c18b36-24da-43b2-8fc1-95a45bef266b" containerName="extract-content" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.527628 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8c18b36-24da-43b2-8fc1-95a45bef266b" containerName="extract-content" Oct 07 14:35:45 crc kubenswrapper[4959]: E1007 14:35:45.527645 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5aa9a4d-4a36-481d-8853-c8a5b586d974" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.527652 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5aa9a4d-4a36-481d-8853-c8a5b586d974" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.527858 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5aa9a4d-4a36-481d-8853-c8a5b586d974" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.527883 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8c18b36-24da-43b2-8fc1-95a45bef266b" containerName="registry-server" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.532248 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.536315 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.536337 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.536712 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.536819 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.536866 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.536962 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-c6vmh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.537035 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.537074 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ceph-nova" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.537442 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.542093 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh"] Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.638574 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.638878 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.639060 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.639265 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.639372 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.639546 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.639621 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.639735 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t47l2\" (UniqueName: \"kubernetes.io/projected/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-kube-api-access-t47l2\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.639821 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.639923 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.639958 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.742426 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.743223 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.743313 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.743349 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.743396 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t47l2\" (UniqueName: \"kubernetes.io/projected/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-kube-api-access-t47l2\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.743432 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.743480 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.743506 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.743542 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.743570 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.743621 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.745131 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.748841 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.749873 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.750022 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.750976 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.752296 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.753475 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.754123 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.757576 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.758683 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.767589 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t47l2\" (UniqueName: \"kubernetes.io/projected/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-kube-api-access-t47l2\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:45 crc kubenswrapper[4959]: I1007 14:35:45.854885 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:35:46 crc kubenswrapper[4959]: I1007 14:35:46.429427 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh"] Oct 07 14:35:47 crc kubenswrapper[4959]: I1007 14:35:47.446309 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" event={"ID":"ce45a1f9-8a26-46d4-a413-c175c4d51ab4","Type":"ContainerStarted","Data":"f6b8541bf27651eca1969fb170711517ca6e8e20e275adcd80a74ac5624ce172"} Oct 07 14:35:47 crc kubenswrapper[4959]: I1007 14:35:47.446659 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" event={"ID":"ce45a1f9-8a26-46d4-a413-c175c4d51ab4","Type":"ContainerStarted","Data":"5e742e338b33e0a19876621abe88b2d0ef82c317e4823ad45f12c934528ecd07"} Oct 07 14:35:47 crc kubenswrapper[4959]: I1007 14:35:47.469415 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" podStartSLOduration=1.804266986 podStartE2EDuration="2.469397738s" podCreationTimestamp="2025-10-07 14:35:45 +0000 UTC" firstStartedPulling="2025-10-07 14:35:46.443710489 +0000 UTC m=+3028.527114814" lastFinishedPulling="2025-10-07 14:35:47.108841241 +0000 UTC m=+3029.192245566" observedRunningTime="2025-10-07 14:35:47.468677969 +0000 UTC m=+3029.552082294" watchObservedRunningTime="2025-10-07 14:35:47.469397738 +0000 UTC m=+3029.552802063" Oct 07 14:35:49 crc kubenswrapper[4959]: I1007 14:35:49.654001 4959 scope.go:117] "RemoveContainer" containerID="5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d" Oct 07 14:35:49 crc kubenswrapper[4959]: E1007 14:35:49.655022 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:36:00 crc kubenswrapper[4959]: I1007 14:36:00.654363 4959 scope.go:117] "RemoveContainer" containerID="5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d" Oct 07 14:36:00 crc kubenswrapper[4959]: E1007 14:36:00.655582 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:36:12 crc kubenswrapper[4959]: I1007 14:36:12.654334 4959 scope.go:117] "RemoveContainer" containerID="5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d" Oct 07 14:36:12 crc kubenswrapper[4959]: E1007 14:36:12.655264 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:36:24 crc kubenswrapper[4959]: I1007 14:36:24.653927 4959 scope.go:117] "RemoveContainer" containerID="5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d" Oct 07 14:36:24 crc kubenswrapper[4959]: E1007 14:36:24.654767 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:36:38 crc kubenswrapper[4959]: I1007 14:36:38.658658 4959 scope.go:117] "RemoveContainer" containerID="5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d" Oct 07 14:36:38 crc kubenswrapper[4959]: E1007 14:36:38.659592 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:36:52 crc kubenswrapper[4959]: I1007 14:36:52.653794 4959 scope.go:117] "RemoveContainer" containerID="5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d" Oct 07 14:36:52 crc kubenswrapper[4959]: E1007 14:36:52.654539 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:37:04 crc kubenswrapper[4959]: I1007 14:37:04.653552 4959 scope.go:117] "RemoveContainer" containerID="5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d" Oct 07 14:37:04 crc kubenswrapper[4959]: E1007 14:37:04.655724 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:37:16 crc kubenswrapper[4959]: I1007 14:37:16.654567 4959 scope.go:117] "RemoveContainer" containerID="5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d" Oct 07 14:37:16 crc kubenswrapper[4959]: E1007 14:37:16.655970 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:37:30 crc kubenswrapper[4959]: I1007 14:37:30.653395 4959 scope.go:117] "RemoveContainer" containerID="5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d" Oct 07 14:37:30 crc kubenswrapper[4959]: E1007 14:37:30.654274 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:37:43 crc kubenswrapper[4959]: I1007 14:37:43.653152 4959 scope.go:117] "RemoveContainer" containerID="5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d" Oct 07 14:37:43 crc kubenswrapper[4959]: E1007 14:37:43.653906 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:37:56 crc kubenswrapper[4959]: I1007 14:37:56.653241 4959 scope.go:117] "RemoveContainer" containerID="5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d" Oct 07 14:37:56 crc kubenswrapper[4959]: E1007 14:37:56.654007 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:38:08 crc kubenswrapper[4959]: I1007 14:38:08.660396 4959 scope.go:117] "RemoveContainer" containerID="5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d" Oct 07 14:38:08 crc kubenswrapper[4959]: E1007 14:38:08.661457 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:38:21 crc kubenswrapper[4959]: I1007 14:38:21.653853 4959 scope.go:117] "RemoveContainer" containerID="5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d" Oct 07 14:38:21 crc kubenswrapper[4959]: E1007 14:38:21.654728 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:38:35 crc kubenswrapper[4959]: I1007 14:38:35.654536 4959 scope.go:117] "RemoveContainer" containerID="5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d" Oct 07 14:38:35 crc kubenswrapper[4959]: E1007 14:38:35.657151 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:38:47 crc kubenswrapper[4959]: I1007 14:38:47.654672 4959 scope.go:117] "RemoveContainer" containerID="5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d" Oct 07 14:38:47 crc kubenswrapper[4959]: E1007 14:38:47.655823 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:39:00 crc kubenswrapper[4959]: I1007 14:39:00.653978 4959 scope.go:117] "RemoveContainer" containerID="5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d" Oct 07 14:39:01 crc kubenswrapper[4959]: I1007 14:39:01.250261 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerStarted","Data":"d95ff1c6b77cf13243ed8e843bbaced75a795144cd83d3ae4f932c4a56aa2beb"} Oct 07 14:39:48 crc kubenswrapper[4959]: I1007 14:39:48.688310 4959 generic.go:334] "Generic (PLEG): container finished" podID="ce45a1f9-8a26-46d4-a413-c175c4d51ab4" containerID="f6b8541bf27651eca1969fb170711517ca6e8e20e275adcd80a74ac5624ce172" exitCode=0 Oct 07 14:39:48 crc kubenswrapper[4959]: I1007 14:39:48.688403 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" event={"ID":"ce45a1f9-8a26-46d4-a413-c175c4d51ab4","Type":"ContainerDied","Data":"f6b8541bf27651eca1969fb170711517ca6e8e20e275adcd80a74ac5624ce172"} Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.115453 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.231597 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-ceph\") pod \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.231643 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-ssh-key\") pod \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.231670 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-inventory\") pod \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.231902 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-migration-ssh-key-0\") pod \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.231940 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-cell1-compute-config-1\") pod \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.232019 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-ceph-nova-0\") pod \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.232061 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-extra-config-0\") pod \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.232080 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-custom-ceph-combined-ca-bundle\") pod \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.232139 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t47l2\" (UniqueName: \"kubernetes.io/projected/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-kube-api-access-t47l2\") pod \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.232164 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-cell1-compute-config-0\") pod \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.232229 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-migration-ssh-key-1\") pod \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\" (UID: \"ce45a1f9-8a26-46d4-a413-c175c4d51ab4\") " Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.239767 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-custom-ceph-combined-ca-bundle" (OuterVolumeSpecName: "nova-custom-ceph-combined-ca-bundle") pod "ce45a1f9-8a26-46d4-a413-c175c4d51ab4" (UID: "ce45a1f9-8a26-46d4-a413-c175c4d51ab4"). InnerVolumeSpecName "nova-custom-ceph-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.259557 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-ceph" (OuterVolumeSpecName: "ceph") pod "ce45a1f9-8a26-46d4-a413-c175c4d51ab4" (UID: "ce45a1f9-8a26-46d4-a413-c175c4d51ab4"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.259613 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-kube-api-access-t47l2" (OuterVolumeSpecName: "kube-api-access-t47l2") pod "ce45a1f9-8a26-46d4-a413-c175c4d51ab4" (UID: "ce45a1f9-8a26-46d4-a413-c175c4d51ab4"). InnerVolumeSpecName "kube-api-access-t47l2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.262774 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-ceph-nova-0" (OuterVolumeSpecName: "ceph-nova-0") pod "ce45a1f9-8a26-46d4-a413-c175c4d51ab4" (UID: "ce45a1f9-8a26-46d4-a413-c175c4d51ab4"). InnerVolumeSpecName "ceph-nova-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.266134 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "ce45a1f9-8a26-46d4-a413-c175c4d51ab4" (UID: "ce45a1f9-8a26-46d4-a413-c175c4d51ab4"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.268848 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "ce45a1f9-8a26-46d4-a413-c175c4d51ab4" (UID: "ce45a1f9-8a26-46d4-a413-c175c4d51ab4"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.271820 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "ce45a1f9-8a26-46d4-a413-c175c4d51ab4" (UID: "ce45a1f9-8a26-46d4-a413-c175c4d51ab4"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.273573 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-inventory" (OuterVolumeSpecName: "inventory") pod "ce45a1f9-8a26-46d4-a413-c175c4d51ab4" (UID: "ce45a1f9-8a26-46d4-a413-c175c4d51ab4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.278996 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "ce45a1f9-8a26-46d4-a413-c175c4d51ab4" (UID: "ce45a1f9-8a26-46d4-a413-c175c4d51ab4"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.289778 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "ce45a1f9-8a26-46d4-a413-c175c4d51ab4" (UID: "ce45a1f9-8a26-46d4-a413-c175c4d51ab4"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.292385 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ce45a1f9-8a26-46d4-a413-c175c4d51ab4" (UID: "ce45a1f9-8a26-46d4-a413-c175c4d51ab4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.334513 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t47l2\" (UniqueName: \"kubernetes.io/projected/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-kube-api-access-t47l2\") on node \"crc\" DevicePath \"\"" Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.334547 4959 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.334556 4959 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.334565 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-ceph\") on node \"crc\" DevicePath \"\"" Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.334574 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.334582 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-inventory\") on node \"crc\" DevicePath \"\"" Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.334591 4959 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.334599 4959 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.334607 4959 reconciler_common.go:293] "Volume detached for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-ceph-nova-0\") on node \"crc\" DevicePath \"\"" Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.334616 4959 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.334625 4959 reconciler_common.go:293] "Volume detached for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce45a1f9-8a26-46d4-a413-c175c4d51ab4-nova-custom-ceph-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.726850 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" event={"ID":"ce45a1f9-8a26-46d4-a413-c175c4d51ab4","Type":"ContainerDied","Data":"5e742e338b33e0a19876621abe88b2d0ef82c317e4823ad45f12c934528ecd07"} Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.726897 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e742e338b33e0a19876621abe88b2d0ef82c317e4823ad45f12c934528ecd07" Oct 07 14:39:50 crc kubenswrapper[4959]: I1007 14:39:50.726918 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.012847 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Oct 07 14:40:05 crc kubenswrapper[4959]: E1007 14:40:05.013896 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce45a1f9-8a26-46d4-a413-c175c4d51ab4" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.013915 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce45a1f9-8a26-46d4-a413-c175c4d51ab4" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.014093 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce45a1f9-8a26-46d4-a413-c175c4d51ab4" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.027339 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.027523 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.030588 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.033522 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.033708 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.033723 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.034911 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.060365 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.158811 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-sys\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.158856 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.158882 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04f1296f-0cd0-4189-9f0d-81da20535c37-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.158899 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-dev\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.158918 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04f1296f-0cd0-4189-9f0d-81da20535c37-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.158942 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-etc-nvme\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.159014 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-run\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.159048 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-sys\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.159065 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-lib-modules\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.159090 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.159136 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef75c9c7-34ea-49a0-94f2-9182065a05c0-scripts\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.159156 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.159170 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.159193 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74qdw\" (UniqueName: \"kubernetes.io/projected/ef75c9c7-34ea-49a0-94f2-9182065a05c0-kube-api-access-74qdw\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.159212 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef75c9c7-34ea-49a0-94f2-9182065a05c0-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.159320 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/04f1296f-0cd0-4189-9f0d-81da20535c37-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.159344 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.159361 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxbtl\" (UniqueName: \"kubernetes.io/projected/04f1296f-0cd0-4189-9f0d-81da20535c37-kube-api-access-cxbtl\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.159376 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ef75c9c7-34ea-49a0-94f2-9182065a05c0-ceph\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.159393 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/04f1296f-0cd0-4189-9f0d-81da20535c37-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.159587 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef75c9c7-34ea-49a0-94f2-9182065a05c0-config-data\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.159634 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-dev\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.159670 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04f1296f-0cd0-4189-9f0d-81da20535c37-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.159695 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-run\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.159721 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.159748 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ef75c9c7-34ea-49a0-94f2-9182065a05c0-config-data-custom\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.159786 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.159821 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.159844 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.159865 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.159888 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.159907 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.262718 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.262784 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-sys\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.262803 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.262825 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04f1296f-0cd0-4189-9f0d-81da20535c37-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.262852 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-dev\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.262873 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04f1296f-0cd0-4189-9f0d-81da20535c37-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.262868 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.262898 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-etc-nvme\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263012 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-run\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263044 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-sys\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263077 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-sys\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263119 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-sys\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263141 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-run\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263151 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-etc-nvme\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263168 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-lib-modules\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263171 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263232 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263267 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-lib-modules\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263285 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef75c9c7-34ea-49a0-94f2-9182065a05c0-scripts\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263320 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263328 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263357 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263391 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263415 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74qdw\" (UniqueName: \"kubernetes.io/projected/ef75c9c7-34ea-49a0-94f2-9182065a05c0-kube-api-access-74qdw\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263445 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef75c9c7-34ea-49a0-94f2-9182065a05c0-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263482 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/04f1296f-0cd0-4189-9f0d-81da20535c37-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263534 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263556 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxbtl\" (UniqueName: \"kubernetes.io/projected/04f1296f-0cd0-4189-9f0d-81da20535c37-kube-api-access-cxbtl\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263579 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ef75c9c7-34ea-49a0-94f2-9182065a05c0-ceph\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263601 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263606 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/04f1296f-0cd0-4189-9f0d-81da20535c37-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263647 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263735 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef75c9c7-34ea-49a0-94f2-9182065a05c0-config-data\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263758 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-dev\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263797 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04f1296f-0cd0-4189-9f0d-81da20535c37-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263824 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-run\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263846 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263869 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ef75c9c7-34ea-49a0-94f2-9182065a05c0-config-data-custom\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263906 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263950 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.263982 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.264024 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.264067 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.264206 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-dev\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.264262 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.264292 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-run\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.264333 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.264486 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.264546 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.264590 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-dev\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.265360 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/04f1296f-0cd0-4189-9f0d-81da20535c37-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.265409 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef75c9c7-34ea-49a0-94f2-9182065a05c0-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.275002 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ef75c9c7-34ea-49a0-94f2-9182065a05c0-config-data-custom\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.278926 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef75c9c7-34ea-49a0-94f2-9182065a05c0-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.279811 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef75c9c7-34ea-49a0-94f2-9182065a05c0-scripts\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.280194 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04f1296f-0cd0-4189-9f0d-81da20535c37-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.280319 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/04f1296f-0cd0-4189-9f0d-81da20535c37-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.280721 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04f1296f-0cd0-4189-9f0d-81da20535c37-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.280799 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04f1296f-0cd0-4189-9f0d-81da20535c37-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.282721 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ef75c9c7-34ea-49a0-94f2-9182065a05c0-ceph\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.288325 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef75c9c7-34ea-49a0-94f2-9182065a05c0-config-data\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.289651 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/04f1296f-0cd0-4189-9f0d-81da20535c37-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.295853 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxbtl\" (UniqueName: \"kubernetes.io/projected/04f1296f-0cd0-4189-9f0d-81da20535c37-kube-api-access-cxbtl\") pod \"cinder-volume-volume1-0\" (UID: \"04f1296f-0cd0-4189-9f0d-81da20535c37\") " pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.296988 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74qdw\" (UniqueName: \"kubernetes.io/projected/ef75c9c7-34ea-49a0-94f2-9182065a05c0-kube-api-access-74qdw\") pod \"cinder-backup-0\" (UID: \"ef75c9c7-34ea-49a0-94f2-9182065a05c0\") " pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.364673 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.382706 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.683187 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-vnn8v"] Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.684559 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-vnn8v" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.693015 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-vnn8v"] Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.776579 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtnl9\" (UniqueName: \"kubernetes.io/projected/4af5827e-e3e8-41ce-b57f-1ef247587194-kube-api-access-qtnl9\") pod \"manila-db-create-vnn8v\" (UID: \"4af5827e-e3e8-41ce-b57f-1ef247587194\") " pod="openstack/manila-db-create-vnn8v" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.859629 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.864891 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.867177 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.868447 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.868615 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-bhw4d" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.868769 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.872140 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.880752 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtnl9\" (UniqueName: \"kubernetes.io/projected/4af5827e-e3e8-41ce-b57f-1ef247587194-kube-api-access-qtnl9\") pod \"manila-db-create-vnn8v\" (UID: \"4af5827e-e3e8-41ce-b57f-1ef247587194\") " pod="openstack/manila-db-create-vnn8v" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.913973 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtnl9\" (UniqueName: \"kubernetes.io/projected/4af5827e-e3e8-41ce-b57f-1ef247587194-kube-api-access-qtnl9\") pod \"manila-db-create-vnn8v\" (UID: \"4af5827e-e3e8-41ce-b57f-1ef247587194\") " pod="openstack/manila-db-create-vnn8v" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.933722 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.935634 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.939409 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.939612 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.943290 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.982468 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/43b9df1d-3d35-45f9-bbad-f39f9d33c1db-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"43b9df1d-3d35-45f9-bbad-f39f9d33c1db\") " pod="openstack/glance-default-external-api-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.982522 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43b9df1d-3d35-45f9-bbad-f39f9d33c1db-config-data\") pod \"glance-default-external-api-0\" (UID: \"43b9df1d-3d35-45f9-bbad-f39f9d33c1db\") " pod="openstack/glance-default-external-api-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.982558 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43b9df1d-3d35-45f9-bbad-f39f9d33c1db-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"43b9df1d-3d35-45f9-bbad-f39f9d33c1db\") " pod="openstack/glance-default-external-api-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.982803 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h46v\" (UniqueName: \"kubernetes.io/projected/43b9df1d-3d35-45f9-bbad-f39f9d33c1db-kube-api-access-4h46v\") pod \"glance-default-external-api-0\" (UID: \"43b9df1d-3d35-45f9-bbad-f39f9d33c1db\") " pod="openstack/glance-default-external-api-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.982876 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/43b9df1d-3d35-45f9-bbad-f39f9d33c1db-ceph\") pod \"glance-default-external-api-0\" (UID: \"43b9df1d-3d35-45f9-bbad-f39f9d33c1db\") " pod="openstack/glance-default-external-api-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.983138 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43b9df1d-3d35-45f9-bbad-f39f9d33c1db-scripts\") pod \"glance-default-external-api-0\" (UID: \"43b9df1d-3d35-45f9-bbad-f39f9d33c1db\") " pod="openstack/glance-default-external-api-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.983178 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"43b9df1d-3d35-45f9-bbad-f39f9d33c1db\") " pod="openstack/glance-default-external-api-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.983219 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/43b9df1d-3d35-45f9-bbad-f39f9d33c1db-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"43b9df1d-3d35-45f9-bbad-f39f9d33c1db\") " pod="openstack/glance-default-external-api-0" Oct 07 14:40:05 crc kubenswrapper[4959]: I1007 14:40:05.983381 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43b9df1d-3d35-45f9-bbad-f39f9d33c1db-logs\") pod \"glance-default-external-api-0\" (UID: \"43b9df1d-3d35-45f9-bbad-f39f9d33c1db\") " pod="openstack/glance-default-external-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.017660 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-vnn8v" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.085031 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14f63797-4c08-4844-a4e6-7075dffc801c-logs\") pod \"glance-default-internal-api-0\" (UID: \"14f63797-4c08-4844-a4e6-7075dffc801c\") " pod="openstack/glance-default-internal-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.085085 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/14f63797-4c08-4844-a4e6-7075dffc801c-ceph\") pod \"glance-default-internal-api-0\" (UID: \"14f63797-4c08-4844-a4e6-7075dffc801c\") " pod="openstack/glance-default-internal-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.085188 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14f63797-4c08-4844-a4e6-7075dffc801c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"14f63797-4c08-4844-a4e6-7075dffc801c\") " pod="openstack/glance-default-internal-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.085245 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43b9df1d-3d35-45f9-bbad-f39f9d33c1db-scripts\") pod \"glance-default-external-api-0\" (UID: \"43b9df1d-3d35-45f9-bbad-f39f9d33c1db\") " pod="openstack/glance-default-external-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.085314 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"43b9df1d-3d35-45f9-bbad-f39f9d33c1db\") " pod="openstack/glance-default-external-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.085990 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14f63797-4c08-4844-a4e6-7075dffc801c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"14f63797-4c08-4844-a4e6-7075dffc801c\") " pod="openstack/glance-default-internal-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.086043 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/43b9df1d-3d35-45f9-bbad-f39f9d33c1db-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"43b9df1d-3d35-45f9-bbad-f39f9d33c1db\") " pod="openstack/glance-default-external-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.086071 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp84v\" (UniqueName: \"kubernetes.io/projected/14f63797-4c08-4844-a4e6-7075dffc801c-kube-api-access-zp84v\") pod \"glance-default-internal-api-0\" (UID: \"14f63797-4c08-4844-a4e6-7075dffc801c\") " pod="openstack/glance-default-internal-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.086233 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43b9df1d-3d35-45f9-bbad-f39f9d33c1db-logs\") pod \"glance-default-external-api-0\" (UID: \"43b9df1d-3d35-45f9-bbad-f39f9d33c1db\") " pod="openstack/glance-default-external-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.086295 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/43b9df1d-3d35-45f9-bbad-f39f9d33c1db-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"43b9df1d-3d35-45f9-bbad-f39f9d33c1db\") " pod="openstack/glance-default-external-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.086310 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"43b9df1d-3d35-45f9-bbad-f39f9d33c1db\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.086327 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43b9df1d-3d35-45f9-bbad-f39f9d33c1db-config-data\") pod \"glance-default-external-api-0\" (UID: \"43b9df1d-3d35-45f9-bbad-f39f9d33c1db\") " pod="openstack/glance-default-external-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.092357 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43b9df1d-3d35-45f9-bbad-f39f9d33c1db-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"43b9df1d-3d35-45f9-bbad-f39f9d33c1db\") " pod="openstack/glance-default-external-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.092418 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14f63797-4c08-4844-a4e6-7075dffc801c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"14f63797-4c08-4844-a4e6-7075dffc801c\") " pod="openstack/glance-default-internal-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.092508 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/43b9df1d-3d35-45f9-bbad-f39f9d33c1db-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"43b9df1d-3d35-45f9-bbad-f39f9d33c1db\") " pod="openstack/glance-default-external-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.092545 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"14f63797-4c08-4844-a4e6-7075dffc801c\") " pod="openstack/glance-default-internal-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.092592 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h46v\" (UniqueName: \"kubernetes.io/projected/43b9df1d-3d35-45f9-bbad-f39f9d33c1db-kube-api-access-4h46v\") pod \"glance-default-external-api-0\" (UID: \"43b9df1d-3d35-45f9-bbad-f39f9d33c1db\") " pod="openstack/glance-default-external-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.092609 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43b9df1d-3d35-45f9-bbad-f39f9d33c1db-logs\") pod \"glance-default-external-api-0\" (UID: \"43b9df1d-3d35-45f9-bbad-f39f9d33c1db\") " pod="openstack/glance-default-external-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.092633 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14f63797-4c08-4844-a4e6-7075dffc801c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"14f63797-4c08-4844-a4e6-7075dffc801c\") " pod="openstack/glance-default-internal-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.092702 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/43b9df1d-3d35-45f9-bbad-f39f9d33c1db-ceph\") pod \"glance-default-external-api-0\" (UID: \"43b9df1d-3d35-45f9-bbad-f39f9d33c1db\") " pod="openstack/glance-default-external-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.092793 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14f63797-4c08-4844-a4e6-7075dffc801c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"14f63797-4c08-4844-a4e6-7075dffc801c\") " pod="openstack/glance-default-internal-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.094261 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43b9df1d-3d35-45f9-bbad-f39f9d33c1db-config-data\") pod \"glance-default-external-api-0\" (UID: \"43b9df1d-3d35-45f9-bbad-f39f9d33c1db\") " pod="openstack/glance-default-external-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.097995 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/43b9df1d-3d35-45f9-bbad-f39f9d33c1db-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"43b9df1d-3d35-45f9-bbad-f39f9d33c1db\") " pod="openstack/glance-default-external-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.100133 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43b9df1d-3d35-45f9-bbad-f39f9d33c1db-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"43b9df1d-3d35-45f9-bbad-f39f9d33c1db\") " pod="openstack/glance-default-external-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.100239 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43b9df1d-3d35-45f9-bbad-f39f9d33c1db-scripts\") pod \"glance-default-external-api-0\" (UID: \"43b9df1d-3d35-45f9-bbad-f39f9d33c1db\") " pod="openstack/glance-default-external-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.100954 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/43b9df1d-3d35-45f9-bbad-f39f9d33c1db-ceph\") pod \"glance-default-external-api-0\" (UID: \"43b9df1d-3d35-45f9-bbad-f39f9d33c1db\") " pod="openstack/glance-default-external-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.112117 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h46v\" (UniqueName: \"kubernetes.io/projected/43b9df1d-3d35-45f9-bbad-f39f9d33c1db-kube-api-access-4h46v\") pod \"glance-default-external-api-0\" (UID: \"43b9df1d-3d35-45f9-bbad-f39f9d33c1db\") " pod="openstack/glance-default-external-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.120608 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.135694 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.137189 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"43b9df1d-3d35-45f9-bbad-f39f9d33c1db\") " pod="openstack/glance-default-external-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.195084 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14f63797-4c08-4844-a4e6-7075dffc801c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"14f63797-4c08-4844-a4e6-7075dffc801c\") " pod="openstack/glance-default-internal-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.195573 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"14f63797-4c08-4844-a4e6-7075dffc801c\") " pod="openstack/glance-default-internal-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.195776 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"14f63797-4c08-4844-a4e6-7075dffc801c\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.196870 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14f63797-4c08-4844-a4e6-7075dffc801c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"14f63797-4c08-4844-a4e6-7075dffc801c\") " pod="openstack/glance-default-internal-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.195608 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/14f63797-4c08-4844-a4e6-7075dffc801c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"14f63797-4c08-4844-a4e6-7075dffc801c\") " pod="openstack/glance-default-internal-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.199376 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14f63797-4c08-4844-a4e6-7075dffc801c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"14f63797-4c08-4844-a4e6-7075dffc801c\") " pod="openstack/glance-default-internal-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.199474 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14f63797-4c08-4844-a4e6-7075dffc801c-logs\") pod \"glance-default-internal-api-0\" (UID: \"14f63797-4c08-4844-a4e6-7075dffc801c\") " pod="openstack/glance-default-internal-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.199496 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/14f63797-4c08-4844-a4e6-7075dffc801c-ceph\") pod \"glance-default-internal-api-0\" (UID: \"14f63797-4c08-4844-a4e6-7075dffc801c\") " pod="openstack/glance-default-internal-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.201328 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14f63797-4c08-4844-a4e6-7075dffc801c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"14f63797-4c08-4844-a4e6-7075dffc801c\") " pod="openstack/glance-default-internal-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.206483 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14f63797-4c08-4844-a4e6-7075dffc801c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"14f63797-4c08-4844-a4e6-7075dffc801c\") " pod="openstack/glance-default-internal-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.211647 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14f63797-4c08-4844-a4e6-7075dffc801c-logs\") pod \"glance-default-internal-api-0\" (UID: \"14f63797-4c08-4844-a4e6-7075dffc801c\") " pod="openstack/glance-default-internal-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.211781 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14f63797-4c08-4844-a4e6-7075dffc801c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"14f63797-4c08-4844-a4e6-7075dffc801c\") " pod="openstack/glance-default-internal-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.211847 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14f63797-4c08-4844-a4e6-7075dffc801c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"14f63797-4c08-4844-a4e6-7075dffc801c\") " pod="openstack/glance-default-internal-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.211900 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp84v\" (UniqueName: \"kubernetes.io/projected/14f63797-4c08-4844-a4e6-7075dffc801c-kube-api-access-zp84v\") pod \"glance-default-internal-api-0\" (UID: \"14f63797-4c08-4844-a4e6-7075dffc801c\") " pod="openstack/glance-default-internal-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.215274 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.224524 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14f63797-4c08-4844-a4e6-7075dffc801c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"14f63797-4c08-4844-a4e6-7075dffc801c\") " pod="openstack/glance-default-internal-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.225091 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14f63797-4c08-4844-a4e6-7075dffc801c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"14f63797-4c08-4844-a4e6-7075dffc801c\") " pod="openstack/glance-default-internal-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.225199 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/14f63797-4c08-4844-a4e6-7075dffc801c-ceph\") pod \"glance-default-internal-api-0\" (UID: \"14f63797-4c08-4844-a4e6-7075dffc801c\") " pod="openstack/glance-default-internal-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.231697 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp84v\" (UniqueName: \"kubernetes.io/projected/14f63797-4c08-4844-a4e6-7075dffc801c-kube-api-access-zp84v\") pod \"glance-default-internal-api-0\" (UID: \"14f63797-4c08-4844-a4e6-7075dffc801c\") " pod="openstack/glance-default-internal-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.250738 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"14f63797-4c08-4844-a4e6-7075dffc801c\") " pod="openstack/glance-default-internal-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.490999 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-vnn8v"] Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.554902 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.811435 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.890960 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"43b9df1d-3d35-45f9-bbad-f39f9d33c1db","Type":"ContainerStarted","Data":"9e3efbcba338dd427091c547ce7a200bbef4799f5e3bc59ba2508aedcdf98e14"} Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.895848 4959 generic.go:334] "Generic (PLEG): container finished" podID="4af5827e-e3e8-41ce-b57f-1ef247587194" containerID="c93a6ebd506351259bbe55103f8ce272d02d62e872c20fb12b4fc2e3cbeaa341" exitCode=0 Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.895931 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-vnn8v" event={"ID":"4af5827e-e3e8-41ce-b57f-1ef247587194","Type":"ContainerDied","Data":"c93a6ebd506351259bbe55103f8ce272d02d62e872c20fb12b4fc2e3cbeaa341"} Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.895964 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-vnn8v" event={"ID":"4af5827e-e3e8-41ce-b57f-1ef247587194","Type":"ContainerStarted","Data":"702b31588ae2c626981f19f47e2f9afb36ed25e09f311768e49c6a1ebdf34b09"} Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.901010 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"ef75c9c7-34ea-49a0-94f2-9182065a05c0","Type":"ContainerStarted","Data":"023a0a1a577c6730528babe64f56cae14a7a453750a32bd8b8970229ee326a36"} Oct 07 14:40:06 crc kubenswrapper[4959]: I1007 14:40:06.976823 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 07 14:40:07 crc kubenswrapper[4959]: I1007 14:40:07.144708 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 07 14:40:07 crc kubenswrapper[4959]: W1007 14:40:07.221172 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod14f63797_4c08_4844_a4e6_7075dffc801c.slice/crio-131a9131408e1b0a5a0b8a95e2b52a9b1815ed2cc404b89e7f22babc9673a7a2 WatchSource:0}: Error finding container 131a9131408e1b0a5a0b8a95e2b52a9b1815ed2cc404b89e7f22babc9673a7a2: Status 404 returned error can't find the container with id 131a9131408e1b0a5a0b8a95e2b52a9b1815ed2cc404b89e7f22babc9673a7a2 Oct 07 14:40:07 crc kubenswrapper[4959]: I1007 14:40:07.924595 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"ef75c9c7-34ea-49a0-94f2-9182065a05c0","Type":"ContainerStarted","Data":"004063e0ad243695e24a3f17eda972de91556896fd97b98e4cdf448cefed6048"} Oct 07 14:40:07 crc kubenswrapper[4959]: I1007 14:40:07.928316 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"43b9df1d-3d35-45f9-bbad-f39f9d33c1db","Type":"ContainerStarted","Data":"18ef3b77503622e2a48b1e72e4bdcab4f877a429e24c4b7e2612a75847298955"} Oct 07 14:40:07 crc kubenswrapper[4959]: I1007 14:40:07.933111 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"04f1296f-0cd0-4189-9f0d-81da20535c37","Type":"ContainerStarted","Data":"55e36be8b9422f2161e684ffade613e06a29298fd59d7fb05a9b4a9f02ddec5e"} Oct 07 14:40:07 crc kubenswrapper[4959]: I1007 14:40:07.934744 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"14f63797-4c08-4844-a4e6-7075dffc801c","Type":"ContainerStarted","Data":"131a9131408e1b0a5a0b8a95e2b52a9b1815ed2cc404b89e7f22babc9673a7a2"} Oct 07 14:40:08 crc kubenswrapper[4959]: I1007 14:40:08.228168 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-vnn8v" Oct 07 14:40:08 crc kubenswrapper[4959]: I1007 14:40:08.370012 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtnl9\" (UniqueName: \"kubernetes.io/projected/4af5827e-e3e8-41ce-b57f-1ef247587194-kube-api-access-qtnl9\") pod \"4af5827e-e3e8-41ce-b57f-1ef247587194\" (UID: \"4af5827e-e3e8-41ce-b57f-1ef247587194\") " Oct 07 14:40:08 crc kubenswrapper[4959]: I1007 14:40:08.375472 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4af5827e-e3e8-41ce-b57f-1ef247587194-kube-api-access-qtnl9" (OuterVolumeSpecName: "kube-api-access-qtnl9") pod "4af5827e-e3e8-41ce-b57f-1ef247587194" (UID: "4af5827e-e3e8-41ce-b57f-1ef247587194"). InnerVolumeSpecName "kube-api-access-qtnl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:40:08 crc kubenswrapper[4959]: I1007 14:40:08.472521 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtnl9\" (UniqueName: \"kubernetes.io/projected/4af5827e-e3e8-41ce-b57f-1ef247587194-kube-api-access-qtnl9\") on node \"crc\" DevicePath \"\"" Oct 07 14:40:08 crc kubenswrapper[4959]: I1007 14:40:08.946743 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"04f1296f-0cd0-4189-9f0d-81da20535c37","Type":"ContainerStarted","Data":"a0fdcb7970274bb100dd17751327061875da068cb842038e50dbb5a75be56489"} Oct 07 14:40:08 crc kubenswrapper[4959]: I1007 14:40:08.949399 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-vnn8v" event={"ID":"4af5827e-e3e8-41ce-b57f-1ef247587194","Type":"ContainerDied","Data":"702b31588ae2c626981f19f47e2f9afb36ed25e09f311768e49c6a1ebdf34b09"} Oct 07 14:40:08 crc kubenswrapper[4959]: I1007 14:40:08.949462 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="702b31588ae2c626981f19f47e2f9afb36ed25e09f311768e49c6a1ebdf34b09" Oct 07 14:40:08 crc kubenswrapper[4959]: I1007 14:40:08.949418 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-vnn8v" Oct 07 14:40:08 crc kubenswrapper[4959]: I1007 14:40:08.951612 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"14f63797-4c08-4844-a4e6-7075dffc801c","Type":"ContainerStarted","Data":"437230f0c012befd64c25dc426a45e6ec098610df76dc49474d6f4d71d99bebe"} Oct 07 14:40:08 crc kubenswrapper[4959]: I1007 14:40:08.954682 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"ef75c9c7-34ea-49a0-94f2-9182065a05c0","Type":"ContainerStarted","Data":"49e81aac46fdbef15e51b1c1bf9c9c7cb63934efabed221d1009877909158895"} Oct 07 14:40:08 crc kubenswrapper[4959]: I1007 14:40:08.956904 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"43b9df1d-3d35-45f9-bbad-f39f9d33c1db","Type":"ContainerStarted","Data":"5c967aa553df5692ab1c56066ef4fecd8b64ad8d08423d093bce4e56d9ae736c"} Oct 07 14:40:08 crc kubenswrapper[4959]: I1007 14:40:08.981618 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.981596683 podStartE2EDuration="4.981596683s" podCreationTimestamp="2025-10-07 14:40:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:40:08.976263937 +0000 UTC m=+3291.059668262" watchObservedRunningTime="2025-10-07 14:40:08.981596683 +0000 UTC m=+3291.065001008" Oct 07 14:40:09 crc kubenswrapper[4959]: I1007 14:40:09.007701 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=3.83248795 podStartE2EDuration="5.007679135s" podCreationTimestamp="2025-10-07 14:40:04 +0000 UTC" firstStartedPulling="2025-10-07 14:40:06.120358658 +0000 UTC m=+3288.203762973" lastFinishedPulling="2025-10-07 14:40:07.295549833 +0000 UTC m=+3289.378954158" observedRunningTime="2025-10-07 14:40:09.000932364 +0000 UTC m=+3291.084336699" watchObservedRunningTime="2025-10-07 14:40:09.007679135 +0000 UTC m=+3291.091083460" Oct 07 14:40:09 crc kubenswrapper[4959]: I1007 14:40:09.967671 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"14f63797-4c08-4844-a4e6-7075dffc801c","Type":"ContainerStarted","Data":"8f95c790054e597d632ba11e20526a1db8af8d288d5bae89f79a97b81c39d347"} Oct 07 14:40:09 crc kubenswrapper[4959]: I1007 14:40:09.971112 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"04f1296f-0cd0-4189-9f0d-81da20535c37","Type":"ContainerStarted","Data":"b870c8769c99666eb94863d0a89689c3eb629f20e4efc21c25c2018da82a53b1"} Oct 07 14:40:09 crc kubenswrapper[4959]: I1007 14:40:09.998129 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.998086168 podStartE2EDuration="5.998086168s" podCreationTimestamp="2025-10-07 14:40:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:40:09.995236086 +0000 UTC m=+3292.078640431" watchObservedRunningTime="2025-10-07 14:40:09.998086168 +0000 UTC m=+3292.081490503" Oct 07 14:40:10 crc kubenswrapper[4959]: I1007 14:40:10.019404 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=3.836697083 podStartE2EDuration="5.019382349s" podCreationTimestamp="2025-10-07 14:40:05 +0000 UTC" firstStartedPulling="2025-10-07 14:40:06.977541737 +0000 UTC m=+3289.060946062" lastFinishedPulling="2025-10-07 14:40:08.160227003 +0000 UTC m=+3290.243631328" observedRunningTime="2025-10-07 14:40:10.01353037 +0000 UTC m=+3292.096934695" watchObservedRunningTime="2025-10-07 14:40:10.019382349 +0000 UTC m=+3292.102786664" Oct 07 14:40:10 crc kubenswrapper[4959]: I1007 14:40:10.366761 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Oct 07 14:40:10 crc kubenswrapper[4959]: I1007 14:40:10.383783 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:15 crc kubenswrapper[4959]: I1007 14:40:15.622411 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Oct 07 14:40:15 crc kubenswrapper[4959]: I1007 14:40:15.636535 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Oct 07 14:40:15 crc kubenswrapper[4959]: I1007 14:40:15.779284 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-09be-account-create-7dklf"] Oct 07 14:40:15 crc kubenswrapper[4959]: E1007 14:40:15.779957 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4af5827e-e3e8-41ce-b57f-1ef247587194" containerName="mariadb-database-create" Oct 07 14:40:15 crc kubenswrapper[4959]: I1007 14:40:15.779980 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4af5827e-e3e8-41ce-b57f-1ef247587194" containerName="mariadb-database-create" Oct 07 14:40:15 crc kubenswrapper[4959]: I1007 14:40:15.780287 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="4af5827e-e3e8-41ce-b57f-1ef247587194" containerName="mariadb-database-create" Oct 07 14:40:15 crc kubenswrapper[4959]: I1007 14:40:15.781125 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-09be-account-create-7dklf" Oct 07 14:40:15 crc kubenswrapper[4959]: I1007 14:40:15.783606 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Oct 07 14:40:15 crc kubenswrapper[4959]: I1007 14:40:15.788759 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-09be-account-create-7dklf"] Oct 07 14:40:15 crc kubenswrapper[4959]: I1007 14:40:15.840856 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8mhr\" (UniqueName: \"kubernetes.io/projected/cd865e19-4b6b-4c3f-818a-bd93e7b23045-kube-api-access-h8mhr\") pod \"manila-09be-account-create-7dklf\" (UID: \"cd865e19-4b6b-4c3f-818a-bd93e7b23045\") " pod="openstack/manila-09be-account-create-7dklf" Oct 07 14:40:15 crc kubenswrapper[4959]: I1007 14:40:15.942856 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8mhr\" (UniqueName: \"kubernetes.io/projected/cd865e19-4b6b-4c3f-818a-bd93e7b23045-kube-api-access-h8mhr\") pod \"manila-09be-account-create-7dklf\" (UID: \"cd865e19-4b6b-4c3f-818a-bd93e7b23045\") " pod="openstack/manila-09be-account-create-7dklf" Oct 07 14:40:15 crc kubenswrapper[4959]: I1007 14:40:15.964867 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8mhr\" (UniqueName: \"kubernetes.io/projected/cd865e19-4b6b-4c3f-818a-bd93e7b23045-kube-api-access-h8mhr\") pod \"manila-09be-account-create-7dklf\" (UID: \"cd865e19-4b6b-4c3f-818a-bd93e7b23045\") " pod="openstack/manila-09be-account-create-7dklf" Oct 07 14:40:16 crc kubenswrapper[4959]: I1007 14:40:16.120391 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-09be-account-create-7dklf" Oct 07 14:40:16 crc kubenswrapper[4959]: I1007 14:40:16.216965 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 07 14:40:16 crc kubenswrapper[4959]: I1007 14:40:16.217435 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 07 14:40:16 crc kubenswrapper[4959]: I1007 14:40:16.269308 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 07 14:40:16 crc kubenswrapper[4959]: I1007 14:40:16.317965 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 07 14:40:16 crc kubenswrapper[4959]: I1007 14:40:16.555876 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 07 14:40:16 crc kubenswrapper[4959]: I1007 14:40:16.555987 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 07 14:40:16 crc kubenswrapper[4959]: I1007 14:40:16.591480 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 07 14:40:16 crc kubenswrapper[4959]: I1007 14:40:16.601293 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 07 14:40:16 crc kubenswrapper[4959]: I1007 14:40:16.646093 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-09be-account-create-7dklf"] Oct 07 14:40:17 crc kubenswrapper[4959]: I1007 14:40:17.048236 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-09be-account-create-7dklf" event={"ID":"cd865e19-4b6b-4c3f-818a-bd93e7b23045","Type":"ContainerStarted","Data":"dbdec36c647e5432c56d495d0ecde7f51ab8164659ea76c66ca9614967b65478"} Oct 07 14:40:17 crc kubenswrapper[4959]: I1007 14:40:17.049009 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 07 14:40:17 crc kubenswrapper[4959]: I1007 14:40:17.049041 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-09be-account-create-7dklf" event={"ID":"cd865e19-4b6b-4c3f-818a-bd93e7b23045","Type":"ContainerStarted","Data":"7ff7b108966817281b9d79cd78b5b40c9313bb26d60f99d18169a344be9064c7"} Oct 07 14:40:17 crc kubenswrapper[4959]: I1007 14:40:17.049060 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 07 14:40:17 crc kubenswrapper[4959]: I1007 14:40:17.049074 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 07 14:40:17 crc kubenswrapper[4959]: I1007 14:40:17.049083 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 07 14:40:17 crc kubenswrapper[4959]: I1007 14:40:17.067459 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-09be-account-create-7dklf" podStartSLOduration=2.067436503 podStartE2EDuration="2.067436503s" podCreationTimestamp="2025-10-07 14:40:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:40:17.064538799 +0000 UTC m=+3299.147943124" watchObservedRunningTime="2025-10-07 14:40:17.067436503 +0000 UTC m=+3299.150840828" Oct 07 14:40:18 crc kubenswrapper[4959]: I1007 14:40:18.061321 4959 generic.go:334] "Generic (PLEG): container finished" podID="cd865e19-4b6b-4c3f-818a-bd93e7b23045" containerID="dbdec36c647e5432c56d495d0ecde7f51ab8164659ea76c66ca9614967b65478" exitCode=0 Oct 07 14:40:18 crc kubenswrapper[4959]: I1007 14:40:18.063375 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-09be-account-create-7dklf" event={"ID":"cd865e19-4b6b-4c3f-818a-bd93e7b23045","Type":"ContainerDied","Data":"dbdec36c647e5432c56d495d0ecde7f51ab8164659ea76c66ca9614967b65478"} Oct 07 14:40:19 crc kubenswrapper[4959]: I1007 14:40:19.515864 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 07 14:40:19 crc kubenswrapper[4959]: I1007 14:40:19.516676 4959 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 07 14:40:19 crc kubenswrapper[4959]: I1007 14:40:19.571605 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 07 14:40:19 crc kubenswrapper[4959]: I1007 14:40:19.630134 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-09be-account-create-7dklf" Oct 07 14:40:19 crc kubenswrapper[4959]: I1007 14:40:19.754055 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8mhr\" (UniqueName: \"kubernetes.io/projected/cd865e19-4b6b-4c3f-818a-bd93e7b23045-kube-api-access-h8mhr\") pod \"cd865e19-4b6b-4c3f-818a-bd93e7b23045\" (UID: \"cd865e19-4b6b-4c3f-818a-bd93e7b23045\") " Oct 07 14:40:19 crc kubenswrapper[4959]: I1007 14:40:19.763697 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd865e19-4b6b-4c3f-818a-bd93e7b23045-kube-api-access-h8mhr" (OuterVolumeSpecName: "kube-api-access-h8mhr") pod "cd865e19-4b6b-4c3f-818a-bd93e7b23045" (UID: "cd865e19-4b6b-4c3f-818a-bd93e7b23045"). InnerVolumeSpecName "kube-api-access-h8mhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:40:19 crc kubenswrapper[4959]: I1007 14:40:19.856167 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8mhr\" (UniqueName: \"kubernetes.io/projected/cd865e19-4b6b-4c3f-818a-bd93e7b23045-kube-api-access-h8mhr\") on node \"crc\" DevicePath \"\"" Oct 07 14:40:19 crc kubenswrapper[4959]: I1007 14:40:19.953247 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 07 14:40:19 crc kubenswrapper[4959]: I1007 14:40:19.953401 4959 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 07 14:40:19 crc kubenswrapper[4959]: I1007 14:40:19.955368 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 07 14:40:20 crc kubenswrapper[4959]: I1007 14:40:20.135446 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-09be-account-create-7dklf" event={"ID":"cd865e19-4b6b-4c3f-818a-bd93e7b23045","Type":"ContainerDied","Data":"7ff7b108966817281b9d79cd78b5b40c9313bb26d60f99d18169a344be9064c7"} Oct 07 14:40:20 crc kubenswrapper[4959]: I1007 14:40:20.135521 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ff7b108966817281b9d79cd78b5b40c9313bb26d60f99d18169a344be9064c7" Oct 07 14:40:20 crc kubenswrapper[4959]: I1007 14:40:20.135609 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-09be-account-create-7dklf" Oct 07 14:40:21 crc kubenswrapper[4959]: I1007 14:40:21.026911 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-bwnpt"] Oct 07 14:40:21 crc kubenswrapper[4959]: E1007 14:40:21.028023 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd865e19-4b6b-4c3f-818a-bd93e7b23045" containerName="mariadb-account-create" Oct 07 14:40:21 crc kubenswrapper[4959]: I1007 14:40:21.028045 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd865e19-4b6b-4c3f-818a-bd93e7b23045" containerName="mariadb-account-create" Oct 07 14:40:21 crc kubenswrapper[4959]: I1007 14:40:21.028308 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd865e19-4b6b-4c3f-818a-bd93e7b23045" containerName="mariadb-account-create" Oct 07 14:40:21 crc kubenswrapper[4959]: I1007 14:40:21.029521 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-bwnpt" Oct 07 14:40:21 crc kubenswrapper[4959]: I1007 14:40:21.032637 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-vl6jc" Oct 07 14:40:21 crc kubenswrapper[4959]: I1007 14:40:21.033993 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 07 14:40:21 crc kubenswrapper[4959]: I1007 14:40:21.048342 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-bwnpt"] Oct 07 14:40:21 crc kubenswrapper[4959]: I1007 14:40:21.089314 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/b129b17a-32c0-4562-bc84-a1e5ea413176-job-config-data\") pod \"manila-db-sync-bwnpt\" (UID: \"b129b17a-32c0-4562-bc84-a1e5ea413176\") " pod="openstack/manila-db-sync-bwnpt" Oct 07 14:40:21 crc kubenswrapper[4959]: I1007 14:40:21.089450 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwhgc\" (UniqueName: \"kubernetes.io/projected/b129b17a-32c0-4562-bc84-a1e5ea413176-kube-api-access-jwhgc\") pod \"manila-db-sync-bwnpt\" (UID: \"b129b17a-32c0-4562-bc84-a1e5ea413176\") " pod="openstack/manila-db-sync-bwnpt" Oct 07 14:40:21 crc kubenswrapper[4959]: I1007 14:40:21.089581 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b129b17a-32c0-4562-bc84-a1e5ea413176-config-data\") pod \"manila-db-sync-bwnpt\" (UID: \"b129b17a-32c0-4562-bc84-a1e5ea413176\") " pod="openstack/manila-db-sync-bwnpt" Oct 07 14:40:21 crc kubenswrapper[4959]: I1007 14:40:21.089698 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b129b17a-32c0-4562-bc84-a1e5ea413176-combined-ca-bundle\") pod \"manila-db-sync-bwnpt\" (UID: \"b129b17a-32c0-4562-bc84-a1e5ea413176\") " pod="openstack/manila-db-sync-bwnpt" Oct 07 14:40:21 crc kubenswrapper[4959]: I1007 14:40:21.192046 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b129b17a-32c0-4562-bc84-a1e5ea413176-combined-ca-bundle\") pod \"manila-db-sync-bwnpt\" (UID: \"b129b17a-32c0-4562-bc84-a1e5ea413176\") " pod="openstack/manila-db-sync-bwnpt" Oct 07 14:40:21 crc kubenswrapper[4959]: I1007 14:40:21.192255 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/b129b17a-32c0-4562-bc84-a1e5ea413176-job-config-data\") pod \"manila-db-sync-bwnpt\" (UID: \"b129b17a-32c0-4562-bc84-a1e5ea413176\") " pod="openstack/manila-db-sync-bwnpt" Oct 07 14:40:21 crc kubenswrapper[4959]: I1007 14:40:21.192306 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwhgc\" (UniqueName: \"kubernetes.io/projected/b129b17a-32c0-4562-bc84-a1e5ea413176-kube-api-access-jwhgc\") pod \"manila-db-sync-bwnpt\" (UID: \"b129b17a-32c0-4562-bc84-a1e5ea413176\") " pod="openstack/manila-db-sync-bwnpt" Oct 07 14:40:21 crc kubenswrapper[4959]: I1007 14:40:21.192355 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b129b17a-32c0-4562-bc84-a1e5ea413176-config-data\") pod \"manila-db-sync-bwnpt\" (UID: \"b129b17a-32c0-4562-bc84-a1e5ea413176\") " pod="openstack/manila-db-sync-bwnpt" Oct 07 14:40:21 crc kubenswrapper[4959]: I1007 14:40:21.203797 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b129b17a-32c0-4562-bc84-a1e5ea413176-combined-ca-bundle\") pod \"manila-db-sync-bwnpt\" (UID: \"b129b17a-32c0-4562-bc84-a1e5ea413176\") " pod="openstack/manila-db-sync-bwnpt" Oct 07 14:40:21 crc kubenswrapper[4959]: I1007 14:40:21.205813 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/b129b17a-32c0-4562-bc84-a1e5ea413176-job-config-data\") pod \"manila-db-sync-bwnpt\" (UID: \"b129b17a-32c0-4562-bc84-a1e5ea413176\") " pod="openstack/manila-db-sync-bwnpt" Oct 07 14:40:21 crc kubenswrapper[4959]: I1007 14:40:21.206798 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b129b17a-32c0-4562-bc84-a1e5ea413176-config-data\") pod \"manila-db-sync-bwnpt\" (UID: \"b129b17a-32c0-4562-bc84-a1e5ea413176\") " pod="openstack/manila-db-sync-bwnpt" Oct 07 14:40:21 crc kubenswrapper[4959]: I1007 14:40:21.256823 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwhgc\" (UniqueName: \"kubernetes.io/projected/b129b17a-32c0-4562-bc84-a1e5ea413176-kube-api-access-jwhgc\") pod \"manila-db-sync-bwnpt\" (UID: \"b129b17a-32c0-4562-bc84-a1e5ea413176\") " pod="openstack/manila-db-sync-bwnpt" Oct 07 14:40:21 crc kubenswrapper[4959]: I1007 14:40:21.351538 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-bwnpt" Oct 07 14:40:21 crc kubenswrapper[4959]: I1007 14:40:21.841120 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-bwnpt"] Oct 07 14:40:21 crc kubenswrapper[4959]: W1007 14:40:21.843721 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb129b17a_32c0_4562_bc84_a1e5ea413176.slice/crio-dc60812ad7c235e58f66fad05d92857dc2553ef937de31e0c8bca9449717dbb1 WatchSource:0}: Error finding container dc60812ad7c235e58f66fad05d92857dc2553ef937de31e0c8bca9449717dbb1: Status 404 returned error can't find the container with id dc60812ad7c235e58f66fad05d92857dc2553ef937de31e0c8bca9449717dbb1 Oct 07 14:40:22 crc kubenswrapper[4959]: I1007 14:40:22.157474 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-bwnpt" event={"ID":"b129b17a-32c0-4562-bc84-a1e5ea413176","Type":"ContainerStarted","Data":"dc60812ad7c235e58f66fad05d92857dc2553ef937de31e0c8bca9449717dbb1"} Oct 07 14:40:30 crc kubenswrapper[4959]: I1007 14:40:30.267453 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-bwnpt" event={"ID":"b129b17a-32c0-4562-bc84-a1e5ea413176","Type":"ContainerStarted","Data":"1a85dc16574d9b283a79a3b9aee8744ea227cc83da582fea4869ead30e401f5a"} Oct 07 14:40:30 crc kubenswrapper[4959]: I1007 14:40:30.313189 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-bwnpt" podStartSLOduration=2.001179637 podStartE2EDuration="9.313166644s" podCreationTimestamp="2025-10-07 14:40:21 +0000 UTC" firstStartedPulling="2025-10-07 14:40:21.846995115 +0000 UTC m=+3303.930399440" lastFinishedPulling="2025-10-07 14:40:29.158982122 +0000 UTC m=+3311.242386447" observedRunningTime="2025-10-07 14:40:30.305054118 +0000 UTC m=+3312.388458453" watchObservedRunningTime="2025-10-07 14:40:30.313166644 +0000 UTC m=+3312.396570969" Oct 07 14:40:42 crc kubenswrapper[4959]: I1007 14:40:42.390149 4959 generic.go:334] "Generic (PLEG): container finished" podID="b129b17a-32c0-4562-bc84-a1e5ea413176" containerID="1a85dc16574d9b283a79a3b9aee8744ea227cc83da582fea4869ead30e401f5a" exitCode=0 Oct 07 14:40:42 crc kubenswrapper[4959]: I1007 14:40:42.390294 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-bwnpt" event={"ID":"b129b17a-32c0-4562-bc84-a1e5ea413176","Type":"ContainerDied","Data":"1a85dc16574d9b283a79a3b9aee8744ea227cc83da582fea4869ead30e401f5a"} Oct 07 14:40:43 crc kubenswrapper[4959]: I1007 14:40:43.858326 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-bwnpt" Oct 07 14:40:43 crc kubenswrapper[4959]: I1007 14:40:43.961032 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/b129b17a-32c0-4562-bc84-a1e5ea413176-job-config-data\") pod \"b129b17a-32c0-4562-bc84-a1e5ea413176\" (UID: \"b129b17a-32c0-4562-bc84-a1e5ea413176\") " Oct 07 14:40:43 crc kubenswrapper[4959]: I1007 14:40:43.961320 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b129b17a-32c0-4562-bc84-a1e5ea413176-combined-ca-bundle\") pod \"b129b17a-32c0-4562-bc84-a1e5ea413176\" (UID: \"b129b17a-32c0-4562-bc84-a1e5ea413176\") " Oct 07 14:40:43 crc kubenswrapper[4959]: I1007 14:40:43.961411 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwhgc\" (UniqueName: \"kubernetes.io/projected/b129b17a-32c0-4562-bc84-a1e5ea413176-kube-api-access-jwhgc\") pod \"b129b17a-32c0-4562-bc84-a1e5ea413176\" (UID: \"b129b17a-32c0-4562-bc84-a1e5ea413176\") " Oct 07 14:40:43 crc kubenswrapper[4959]: I1007 14:40:43.961581 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b129b17a-32c0-4562-bc84-a1e5ea413176-config-data\") pod \"b129b17a-32c0-4562-bc84-a1e5ea413176\" (UID: \"b129b17a-32c0-4562-bc84-a1e5ea413176\") " Oct 07 14:40:43 crc kubenswrapper[4959]: I1007 14:40:43.970322 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b129b17a-32c0-4562-bc84-a1e5ea413176-kube-api-access-jwhgc" (OuterVolumeSpecName: "kube-api-access-jwhgc") pod "b129b17a-32c0-4562-bc84-a1e5ea413176" (UID: "b129b17a-32c0-4562-bc84-a1e5ea413176"). InnerVolumeSpecName "kube-api-access-jwhgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:40:43 crc kubenswrapper[4959]: I1007 14:40:43.971056 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b129b17a-32c0-4562-bc84-a1e5ea413176-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "b129b17a-32c0-4562-bc84-a1e5ea413176" (UID: "b129b17a-32c0-4562-bc84-a1e5ea413176"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:40:43 crc kubenswrapper[4959]: I1007 14:40:43.973601 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b129b17a-32c0-4562-bc84-a1e5ea413176-config-data" (OuterVolumeSpecName: "config-data") pod "b129b17a-32c0-4562-bc84-a1e5ea413176" (UID: "b129b17a-32c0-4562-bc84-a1e5ea413176"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:40:43 crc kubenswrapper[4959]: I1007 14:40:43.997842 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b129b17a-32c0-4562-bc84-a1e5ea413176-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b129b17a-32c0-4562-bc84-a1e5ea413176" (UID: "b129b17a-32c0-4562-bc84-a1e5ea413176"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:40:44 crc kubenswrapper[4959]: I1007 14:40:44.065821 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwhgc\" (UniqueName: \"kubernetes.io/projected/b129b17a-32c0-4562-bc84-a1e5ea413176-kube-api-access-jwhgc\") on node \"crc\" DevicePath \"\"" Oct 07 14:40:44 crc kubenswrapper[4959]: I1007 14:40:44.065878 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b129b17a-32c0-4562-bc84-a1e5ea413176-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:40:44 crc kubenswrapper[4959]: I1007 14:40:44.065900 4959 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/b129b17a-32c0-4562-bc84-a1e5ea413176-job-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:40:44 crc kubenswrapper[4959]: I1007 14:40:44.065913 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b129b17a-32c0-4562-bc84-a1e5ea413176-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:40:44 crc kubenswrapper[4959]: I1007 14:40:44.419629 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-bwnpt" event={"ID":"b129b17a-32c0-4562-bc84-a1e5ea413176","Type":"ContainerDied","Data":"dc60812ad7c235e58f66fad05d92857dc2553ef937de31e0c8bca9449717dbb1"} Oct 07 14:40:44 crc kubenswrapper[4959]: I1007 14:40:44.419687 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc60812ad7c235e58f66fad05d92857dc2553ef937de31e0c8bca9449717dbb1" Oct 07 14:40:44 crc kubenswrapper[4959]: I1007 14:40:44.419767 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-bwnpt" Oct 07 14:40:44 crc kubenswrapper[4959]: I1007 14:40:44.758009 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Oct 07 14:40:44 crc kubenswrapper[4959]: E1007 14:40:44.759443 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b129b17a-32c0-4562-bc84-a1e5ea413176" containerName="manila-db-sync" Oct 07 14:40:44 crc kubenswrapper[4959]: I1007 14:40:44.759464 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b129b17a-32c0-4562-bc84-a1e5ea413176" containerName="manila-db-sync" Oct 07 14:40:44 crc kubenswrapper[4959]: I1007 14:40:44.760244 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b129b17a-32c0-4562-bc84-a1e5ea413176" containerName="manila-db-sync" Oct 07 14:40:44 crc kubenswrapper[4959]: I1007 14:40:44.762558 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 07 14:40:44 crc kubenswrapper[4959]: I1007 14:40:44.765792 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 07 14:40:44 crc kubenswrapper[4959]: I1007 14:40:44.765883 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-vl6jc" Oct 07 14:40:44 crc kubenswrapper[4959]: I1007 14:40:44.767419 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Oct 07 14:40:44 crc kubenswrapper[4959]: I1007 14:40:44.768767 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Oct 07 14:40:44 crc kubenswrapper[4959]: I1007 14:40:44.782134 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 07 14:40:44 crc kubenswrapper[4959]: I1007 14:40:44.887146 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/92f09b71-3264-41e1-85b9-86ea71cffc91-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"92f09b71-3264-41e1-85b9-86ea71cffc91\") " pod="openstack/manila-scheduler-0" Oct 07 14:40:44 crc kubenswrapper[4959]: I1007 14:40:44.887418 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92f09b71-3264-41e1-85b9-86ea71cffc91-scripts\") pod \"manila-scheduler-0\" (UID: \"92f09b71-3264-41e1-85b9-86ea71cffc91\") " pod="openstack/manila-scheduler-0" Oct 07 14:40:44 crc kubenswrapper[4959]: I1007 14:40:44.887567 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/92f09b71-3264-41e1-85b9-86ea71cffc91-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"92f09b71-3264-41e1-85b9-86ea71cffc91\") " pod="openstack/manila-scheduler-0" Oct 07 14:40:44 crc kubenswrapper[4959]: I1007 14:40:44.887644 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92f09b71-3264-41e1-85b9-86ea71cffc91-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"92f09b71-3264-41e1-85b9-86ea71cffc91\") " pod="openstack/manila-scheduler-0" Oct 07 14:40:44 crc kubenswrapper[4959]: I1007 14:40:44.887673 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftbqx\" (UniqueName: \"kubernetes.io/projected/92f09b71-3264-41e1-85b9-86ea71cffc91-kube-api-access-ftbqx\") pod \"manila-scheduler-0\" (UID: \"92f09b71-3264-41e1-85b9-86ea71cffc91\") " pod="openstack/manila-scheduler-0" Oct 07 14:40:44 crc kubenswrapper[4959]: I1007 14:40:44.887698 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92f09b71-3264-41e1-85b9-86ea71cffc91-config-data\") pod \"manila-scheduler-0\" (UID: \"92f09b71-3264-41e1-85b9-86ea71cffc91\") " pod="openstack/manila-scheduler-0" Oct 07 14:40:44 crc kubenswrapper[4959]: I1007 14:40:44.924952 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Oct 07 14:40:44 crc kubenswrapper[4959]: I1007 14:40:44.939480 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 07 14:40:44 crc kubenswrapper[4959]: I1007 14:40:44.960313 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Oct 07 14:40:44 crc kubenswrapper[4959]: I1007 14:40:44.973545 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79d4ff65f9-n8c9v"] Oct 07 14:40:44 crc kubenswrapper[4959]: I1007 14:40:44.983970 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79d4ff65f9-n8c9v" Oct 07 14:40:44 crc kubenswrapper[4959]: I1007 14:40:44.996186 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/92f09b71-3264-41e1-85b9-86ea71cffc91-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"92f09b71-3264-41e1-85b9-86ea71cffc91\") " pod="openstack/manila-scheduler-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:44.997772 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/92f09b71-3264-41e1-85b9-86ea71cffc91-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"92f09b71-3264-41e1-85b9-86ea71cffc91\") " pod="openstack/manila-scheduler-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.003534 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92f09b71-3264-41e1-85b9-86ea71cffc91-scripts\") pod \"manila-scheduler-0\" (UID: \"92f09b71-3264-41e1-85b9-86ea71cffc91\") " pod="openstack/manila-scheduler-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.003696 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/92f09b71-3264-41e1-85b9-86ea71cffc91-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"92f09b71-3264-41e1-85b9-86ea71cffc91\") " pod="openstack/manila-scheduler-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.003810 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftbqx\" (UniqueName: \"kubernetes.io/projected/92f09b71-3264-41e1-85b9-86ea71cffc91-kube-api-access-ftbqx\") pod \"manila-scheduler-0\" (UID: \"92f09b71-3264-41e1-85b9-86ea71cffc91\") " pod="openstack/manila-scheduler-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.003841 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92f09b71-3264-41e1-85b9-86ea71cffc91-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"92f09b71-3264-41e1-85b9-86ea71cffc91\") " pod="openstack/manila-scheduler-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.003882 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92f09b71-3264-41e1-85b9-86ea71cffc91-config-data\") pod \"manila-scheduler-0\" (UID: \"92f09b71-3264-41e1-85b9-86ea71cffc91\") " pod="openstack/manila-scheduler-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.017148 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92f09b71-3264-41e1-85b9-86ea71cffc91-scripts\") pod \"manila-scheduler-0\" (UID: \"92f09b71-3264-41e1-85b9-86ea71cffc91\") " pod="openstack/manila-scheduler-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.019323 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92f09b71-3264-41e1-85b9-86ea71cffc91-config-data\") pod \"manila-scheduler-0\" (UID: \"92f09b71-3264-41e1-85b9-86ea71cffc91\") " pod="openstack/manila-scheduler-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.020330 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92f09b71-3264-41e1-85b9-86ea71cffc91-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"92f09b71-3264-41e1-85b9-86ea71cffc91\") " pod="openstack/manila-scheduler-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.022000 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/92f09b71-3264-41e1-85b9-86ea71cffc91-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"92f09b71-3264-41e1-85b9-86ea71cffc91\") " pod="openstack/manila-scheduler-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.023217 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.059871 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79d4ff65f9-n8c9v"] Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.073952 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftbqx\" (UniqueName: \"kubernetes.io/projected/92f09b71-3264-41e1-85b9-86ea71cffc91-kube-api-access-ftbqx\") pod \"manila-scheduler-0\" (UID: \"92f09b71-3264-41e1-85b9-86ea71cffc91\") " pod="openstack/manila-scheduler-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.106462 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/418c77d5-2702-4326-93f7-738440fbe84f-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " pod="openstack/manila-share-share1-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.106529 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/418c77d5-2702-4326-93f7-738440fbe84f-config-data\") pod \"manila-share-share1-0\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " pod="openstack/manila-share-share1-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.106563 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d46df3be-9794-4019-8099-9c3757b5b468-config\") pod \"dnsmasq-dns-79d4ff65f9-n8c9v\" (UID: \"d46df3be-9794-4019-8099-9c3757b5b468\") " pod="openstack/dnsmasq-dns-79d4ff65f9-n8c9v" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.106607 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d46df3be-9794-4019-8099-9c3757b5b468-dns-svc\") pod \"dnsmasq-dns-79d4ff65f9-n8c9v\" (UID: \"d46df3be-9794-4019-8099-9c3757b5b468\") " pod="openstack/dnsmasq-dns-79d4ff65f9-n8c9v" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.106638 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqh7w\" (UniqueName: \"kubernetes.io/projected/418c77d5-2702-4326-93f7-738440fbe84f-kube-api-access-bqh7w\") pod \"manila-share-share1-0\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " pod="openstack/manila-share-share1-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.106687 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d46df3be-9794-4019-8099-9c3757b5b468-ovsdbserver-sb\") pod \"dnsmasq-dns-79d4ff65f9-n8c9v\" (UID: \"d46df3be-9794-4019-8099-9c3757b5b468\") " pod="openstack/dnsmasq-dns-79d4ff65f9-n8c9v" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.106707 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/418c77d5-2702-4326-93f7-738440fbe84f-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " pod="openstack/manila-share-share1-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.106740 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d46df3be-9794-4019-8099-9c3757b5b468-openstack-edpm-ipam\") pod \"dnsmasq-dns-79d4ff65f9-n8c9v\" (UID: \"d46df3be-9794-4019-8099-9c3757b5b468\") " pod="openstack/dnsmasq-dns-79d4ff65f9-n8c9v" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.106767 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/418c77d5-2702-4326-93f7-738440fbe84f-ceph\") pod \"manila-share-share1-0\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " pod="openstack/manila-share-share1-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.106792 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/418c77d5-2702-4326-93f7-738440fbe84f-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " pod="openstack/manila-share-share1-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.106818 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d46df3be-9794-4019-8099-9c3757b5b468-ovsdbserver-nb\") pod \"dnsmasq-dns-79d4ff65f9-n8c9v\" (UID: \"d46df3be-9794-4019-8099-9c3757b5b468\") " pod="openstack/dnsmasq-dns-79d4ff65f9-n8c9v" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.106836 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cm8ml\" (UniqueName: \"kubernetes.io/projected/d46df3be-9794-4019-8099-9c3757b5b468-kube-api-access-cm8ml\") pod \"dnsmasq-dns-79d4ff65f9-n8c9v\" (UID: \"d46df3be-9794-4019-8099-9c3757b5b468\") " pod="openstack/dnsmasq-dns-79d4ff65f9-n8c9v" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.106853 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/418c77d5-2702-4326-93f7-738440fbe84f-scripts\") pod \"manila-share-share1-0\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " pod="openstack/manila-share-share1-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.106885 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/418c77d5-2702-4326-93f7-738440fbe84f-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " pod="openstack/manila-share-share1-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.109353 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.162969 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.165120 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.171043 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.182164 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.209040 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d46df3be-9794-4019-8099-9c3757b5b468-ovsdbserver-sb\") pod \"dnsmasq-dns-79d4ff65f9-n8c9v\" (UID: \"d46df3be-9794-4019-8099-9c3757b5b468\") " pod="openstack/dnsmasq-dns-79d4ff65f9-n8c9v" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.209086 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/418c77d5-2702-4326-93f7-738440fbe84f-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " pod="openstack/manila-share-share1-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.209173 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d46df3be-9794-4019-8099-9c3757b5b468-openstack-edpm-ipam\") pod \"dnsmasq-dns-79d4ff65f9-n8c9v\" (UID: \"d46df3be-9794-4019-8099-9c3757b5b468\") " pod="openstack/dnsmasq-dns-79d4ff65f9-n8c9v" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.209201 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/418c77d5-2702-4326-93f7-738440fbe84f-ceph\") pod \"manila-share-share1-0\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " pod="openstack/manila-share-share1-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.209225 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/418c77d5-2702-4326-93f7-738440fbe84f-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " pod="openstack/manila-share-share1-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.209246 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d46df3be-9794-4019-8099-9c3757b5b468-ovsdbserver-nb\") pod \"dnsmasq-dns-79d4ff65f9-n8c9v\" (UID: \"d46df3be-9794-4019-8099-9c3757b5b468\") " pod="openstack/dnsmasq-dns-79d4ff65f9-n8c9v" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.209263 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cm8ml\" (UniqueName: \"kubernetes.io/projected/d46df3be-9794-4019-8099-9c3757b5b468-kube-api-access-cm8ml\") pod \"dnsmasq-dns-79d4ff65f9-n8c9v\" (UID: \"d46df3be-9794-4019-8099-9c3757b5b468\") " pod="openstack/dnsmasq-dns-79d4ff65f9-n8c9v" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.209279 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/418c77d5-2702-4326-93f7-738440fbe84f-scripts\") pod \"manila-share-share1-0\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " pod="openstack/manila-share-share1-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.209308 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/418c77d5-2702-4326-93f7-738440fbe84f-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " pod="openstack/manila-share-share1-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.209360 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/418c77d5-2702-4326-93f7-738440fbe84f-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " pod="openstack/manila-share-share1-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.209382 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/418c77d5-2702-4326-93f7-738440fbe84f-config-data\") pod \"manila-share-share1-0\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " pod="openstack/manila-share-share1-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.209409 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d46df3be-9794-4019-8099-9c3757b5b468-config\") pod \"dnsmasq-dns-79d4ff65f9-n8c9v\" (UID: \"d46df3be-9794-4019-8099-9c3757b5b468\") " pod="openstack/dnsmasq-dns-79d4ff65f9-n8c9v" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.209444 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d46df3be-9794-4019-8099-9c3757b5b468-dns-svc\") pod \"dnsmasq-dns-79d4ff65f9-n8c9v\" (UID: \"d46df3be-9794-4019-8099-9c3757b5b468\") " pod="openstack/dnsmasq-dns-79d4ff65f9-n8c9v" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.209480 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqh7w\" (UniqueName: \"kubernetes.io/projected/418c77d5-2702-4326-93f7-738440fbe84f-kube-api-access-bqh7w\") pod \"manila-share-share1-0\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " pod="openstack/manila-share-share1-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.210574 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/418c77d5-2702-4326-93f7-738440fbe84f-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " pod="openstack/manila-share-share1-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.211219 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d46df3be-9794-4019-8099-9c3757b5b468-ovsdbserver-sb\") pod \"dnsmasq-dns-79d4ff65f9-n8c9v\" (UID: \"d46df3be-9794-4019-8099-9c3757b5b468\") " pod="openstack/dnsmasq-dns-79d4ff65f9-n8c9v" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.211492 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d46df3be-9794-4019-8099-9c3757b5b468-openstack-edpm-ipam\") pod \"dnsmasq-dns-79d4ff65f9-n8c9v\" (UID: \"d46df3be-9794-4019-8099-9c3757b5b468\") " pod="openstack/dnsmasq-dns-79d4ff65f9-n8c9v" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.213592 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/418c77d5-2702-4326-93f7-738440fbe84f-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " pod="openstack/manila-share-share1-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.225323 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d46df3be-9794-4019-8099-9c3757b5b468-ovsdbserver-nb\") pod \"dnsmasq-dns-79d4ff65f9-n8c9v\" (UID: \"d46df3be-9794-4019-8099-9c3757b5b468\") " pod="openstack/dnsmasq-dns-79d4ff65f9-n8c9v" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.225970 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d46df3be-9794-4019-8099-9c3757b5b468-config\") pod \"dnsmasq-dns-79d4ff65f9-n8c9v\" (UID: \"d46df3be-9794-4019-8099-9c3757b5b468\") " pod="openstack/dnsmasq-dns-79d4ff65f9-n8c9v" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.226631 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d46df3be-9794-4019-8099-9c3757b5b468-dns-svc\") pod \"dnsmasq-dns-79d4ff65f9-n8c9v\" (UID: \"d46df3be-9794-4019-8099-9c3757b5b468\") " pod="openstack/dnsmasq-dns-79d4ff65f9-n8c9v" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.228881 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/418c77d5-2702-4326-93f7-738440fbe84f-ceph\") pod \"manila-share-share1-0\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " pod="openstack/manila-share-share1-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.241970 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/418c77d5-2702-4326-93f7-738440fbe84f-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " pod="openstack/manila-share-share1-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.242246 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/418c77d5-2702-4326-93f7-738440fbe84f-config-data\") pod \"manila-share-share1-0\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " pod="openstack/manila-share-share1-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.242487 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/418c77d5-2702-4326-93f7-738440fbe84f-scripts\") pod \"manila-share-share1-0\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " pod="openstack/manila-share-share1-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.243276 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/418c77d5-2702-4326-93f7-738440fbe84f-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " pod="openstack/manila-share-share1-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.253278 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqh7w\" (UniqueName: \"kubernetes.io/projected/418c77d5-2702-4326-93f7-738440fbe84f-kube-api-access-bqh7w\") pod \"manila-share-share1-0\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " pod="openstack/manila-share-share1-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.265618 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cm8ml\" (UniqueName: \"kubernetes.io/projected/d46df3be-9794-4019-8099-9c3757b5b468-kube-api-access-cm8ml\") pod \"dnsmasq-dns-79d4ff65f9-n8c9v\" (UID: \"d46df3be-9794-4019-8099-9c3757b5b468\") " pod="openstack/dnsmasq-dns-79d4ff65f9-n8c9v" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.274122 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.314282 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsggg\" (UniqueName: \"kubernetes.io/projected/78a6326b-c7db-484b-8065-481b55b8297e-kube-api-access-jsggg\") pod \"manila-api-0\" (UID: \"78a6326b-c7db-484b-8065-481b55b8297e\") " pod="openstack/manila-api-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.314323 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78a6326b-c7db-484b-8065-481b55b8297e-logs\") pod \"manila-api-0\" (UID: \"78a6326b-c7db-484b-8065-481b55b8297e\") " pod="openstack/manila-api-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.314427 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78a6326b-c7db-484b-8065-481b55b8297e-config-data-custom\") pod \"manila-api-0\" (UID: \"78a6326b-c7db-484b-8065-481b55b8297e\") " pod="openstack/manila-api-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.314455 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78a6326b-c7db-484b-8065-481b55b8297e-etc-machine-id\") pod \"manila-api-0\" (UID: \"78a6326b-c7db-484b-8065-481b55b8297e\") " pod="openstack/manila-api-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.314479 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78a6326b-c7db-484b-8065-481b55b8297e-scripts\") pod \"manila-api-0\" (UID: \"78a6326b-c7db-484b-8065-481b55b8297e\") " pod="openstack/manila-api-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.314515 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78a6326b-c7db-484b-8065-481b55b8297e-config-data\") pod \"manila-api-0\" (UID: \"78a6326b-c7db-484b-8065-481b55b8297e\") " pod="openstack/manila-api-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.314640 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78a6326b-c7db-484b-8065-481b55b8297e-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"78a6326b-c7db-484b-8065-481b55b8297e\") " pod="openstack/manila-api-0" Oct 07 14:40:45 crc kubenswrapper[4959]: I1007 14:40:45.318350 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79d4ff65f9-n8c9v" Oct 07 14:40:46 crc kubenswrapper[4959]: I1007 14:40:45.420969 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78a6326b-c7db-484b-8065-481b55b8297e-config-data-custom\") pod \"manila-api-0\" (UID: \"78a6326b-c7db-484b-8065-481b55b8297e\") " pod="openstack/manila-api-0" Oct 07 14:40:46 crc kubenswrapper[4959]: I1007 14:40:45.421379 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78a6326b-c7db-484b-8065-481b55b8297e-etc-machine-id\") pod \"manila-api-0\" (UID: \"78a6326b-c7db-484b-8065-481b55b8297e\") " pod="openstack/manila-api-0" Oct 07 14:40:46 crc kubenswrapper[4959]: I1007 14:40:45.422031 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78a6326b-c7db-484b-8065-481b55b8297e-etc-machine-id\") pod \"manila-api-0\" (UID: \"78a6326b-c7db-484b-8065-481b55b8297e\") " pod="openstack/manila-api-0" Oct 07 14:40:46 crc kubenswrapper[4959]: I1007 14:40:45.423525 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78a6326b-c7db-484b-8065-481b55b8297e-scripts\") pod \"manila-api-0\" (UID: \"78a6326b-c7db-484b-8065-481b55b8297e\") " pod="openstack/manila-api-0" Oct 07 14:40:46 crc kubenswrapper[4959]: I1007 14:40:45.423630 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78a6326b-c7db-484b-8065-481b55b8297e-config-data\") pod \"manila-api-0\" (UID: \"78a6326b-c7db-484b-8065-481b55b8297e\") " pod="openstack/manila-api-0" Oct 07 14:40:46 crc kubenswrapper[4959]: I1007 14:40:45.423671 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78a6326b-c7db-484b-8065-481b55b8297e-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"78a6326b-c7db-484b-8065-481b55b8297e\") " pod="openstack/manila-api-0" Oct 07 14:40:46 crc kubenswrapper[4959]: I1007 14:40:45.423789 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsggg\" (UniqueName: \"kubernetes.io/projected/78a6326b-c7db-484b-8065-481b55b8297e-kube-api-access-jsggg\") pod \"manila-api-0\" (UID: \"78a6326b-c7db-484b-8065-481b55b8297e\") " pod="openstack/manila-api-0" Oct 07 14:40:46 crc kubenswrapper[4959]: I1007 14:40:45.423824 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78a6326b-c7db-484b-8065-481b55b8297e-logs\") pod \"manila-api-0\" (UID: \"78a6326b-c7db-484b-8065-481b55b8297e\") " pod="openstack/manila-api-0" Oct 07 14:40:46 crc kubenswrapper[4959]: I1007 14:40:45.424501 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78a6326b-c7db-484b-8065-481b55b8297e-logs\") pod \"manila-api-0\" (UID: \"78a6326b-c7db-484b-8065-481b55b8297e\") " pod="openstack/manila-api-0" Oct 07 14:40:46 crc kubenswrapper[4959]: I1007 14:40:45.429361 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78a6326b-c7db-484b-8065-481b55b8297e-config-data\") pod \"manila-api-0\" (UID: \"78a6326b-c7db-484b-8065-481b55b8297e\") " pod="openstack/manila-api-0" Oct 07 14:40:46 crc kubenswrapper[4959]: I1007 14:40:45.435861 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78a6326b-c7db-484b-8065-481b55b8297e-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"78a6326b-c7db-484b-8065-481b55b8297e\") " pod="openstack/manila-api-0" Oct 07 14:40:46 crc kubenswrapper[4959]: I1007 14:40:45.436371 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78a6326b-c7db-484b-8065-481b55b8297e-config-data-custom\") pod \"manila-api-0\" (UID: \"78a6326b-c7db-484b-8065-481b55b8297e\") " pod="openstack/manila-api-0" Oct 07 14:40:46 crc kubenswrapper[4959]: I1007 14:40:45.436387 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78a6326b-c7db-484b-8065-481b55b8297e-scripts\") pod \"manila-api-0\" (UID: \"78a6326b-c7db-484b-8065-481b55b8297e\") " pod="openstack/manila-api-0" Oct 07 14:40:46 crc kubenswrapper[4959]: I1007 14:40:45.451235 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsggg\" (UniqueName: \"kubernetes.io/projected/78a6326b-c7db-484b-8065-481b55b8297e-kube-api-access-jsggg\") pod \"manila-api-0\" (UID: \"78a6326b-c7db-484b-8065-481b55b8297e\") " pod="openstack/manila-api-0" Oct 07 14:40:46 crc kubenswrapper[4959]: I1007 14:40:45.625585 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 07 14:40:46 crc kubenswrapper[4959]: I1007 14:40:46.454644 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 07 14:40:46 crc kubenswrapper[4959]: W1007 14:40:46.542652 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd46df3be_9794_4019_8099_9c3757b5b468.slice/crio-e9b9b70fcae188a88a88272e494e1085f76fa12491987bbe2f9936484d3199aa WatchSource:0}: Error finding container e9b9b70fcae188a88a88272e494e1085f76fa12491987bbe2f9936484d3199aa: Status 404 returned error can't find the container with id e9b9b70fcae188a88a88272e494e1085f76fa12491987bbe2f9936484d3199aa Oct 07 14:40:46 crc kubenswrapper[4959]: I1007 14:40:46.544276 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79d4ff65f9-n8c9v"] Oct 07 14:40:46 crc kubenswrapper[4959]: I1007 14:40:46.634937 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 07 14:40:46 crc kubenswrapper[4959]: I1007 14:40:46.720157 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 07 14:40:46 crc kubenswrapper[4959]: W1007 14:40:46.726730 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78a6326b_c7db_484b_8065_481b55b8297e.slice/crio-3d448687467142d25d309eb24914c822557f56d6acfd05eb3778c77813abaf06 WatchSource:0}: Error finding container 3d448687467142d25d309eb24914c822557f56d6acfd05eb3778c77813abaf06: Status 404 returned error can't find the container with id 3d448687467142d25d309eb24914c822557f56d6acfd05eb3778c77813abaf06 Oct 07 14:40:47 crc kubenswrapper[4959]: I1007 14:40:47.466676 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"92f09b71-3264-41e1-85b9-86ea71cffc91","Type":"ContainerStarted","Data":"f39351c89b9ac5602da0f78ed2c464720a1b9d7dd07240f0a51c0f5ac9891d88"} Oct 07 14:40:47 crc kubenswrapper[4959]: I1007 14:40:47.471299 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"78a6326b-c7db-484b-8065-481b55b8297e","Type":"ContainerStarted","Data":"be93772201bc7be59a14408abebae07a87490981f194939cf363d98caf98ce6f"} Oct 07 14:40:47 crc kubenswrapper[4959]: I1007 14:40:47.471368 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"78a6326b-c7db-484b-8065-481b55b8297e","Type":"ContainerStarted","Data":"3d448687467142d25d309eb24914c822557f56d6acfd05eb3778c77813abaf06"} Oct 07 14:40:47 crc kubenswrapper[4959]: I1007 14:40:47.472739 4959 generic.go:334] "Generic (PLEG): container finished" podID="d46df3be-9794-4019-8099-9c3757b5b468" containerID="e6742402f058c05477656aafb8e5b0ae3a70d3fba893d9c69815ddf8e9b271d7" exitCode=0 Oct 07 14:40:47 crc kubenswrapper[4959]: I1007 14:40:47.472841 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79d4ff65f9-n8c9v" event={"ID":"d46df3be-9794-4019-8099-9c3757b5b468","Type":"ContainerDied","Data":"e6742402f058c05477656aafb8e5b0ae3a70d3fba893d9c69815ddf8e9b271d7"} Oct 07 14:40:47 crc kubenswrapper[4959]: I1007 14:40:47.472881 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79d4ff65f9-n8c9v" event={"ID":"d46df3be-9794-4019-8099-9c3757b5b468","Type":"ContainerStarted","Data":"e9b9b70fcae188a88a88272e494e1085f76fa12491987bbe2f9936484d3199aa"} Oct 07 14:40:47 crc kubenswrapper[4959]: I1007 14:40:47.473818 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"418c77d5-2702-4326-93f7-738440fbe84f","Type":"ContainerStarted","Data":"5fcff236de49d79074b29fdb1bfcf60a4dc98fa107b6baea8dc4bfb2817ad0e1"} Oct 07 14:40:47 crc kubenswrapper[4959]: I1007 14:40:47.842728 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Oct 07 14:40:48 crc kubenswrapper[4959]: I1007 14:40:48.488968 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"78a6326b-c7db-484b-8065-481b55b8297e","Type":"ContainerStarted","Data":"6a8e05662badada9ae0d190158e03574fe69fd7cb4fc19cce856f6cee03101c7"} Oct 07 14:40:48 crc kubenswrapper[4959]: I1007 14:40:48.493218 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Oct 07 14:40:48 crc kubenswrapper[4959]: I1007 14:40:48.489163 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="78a6326b-c7db-484b-8065-481b55b8297e" containerName="manila-api" containerID="cri-o://6a8e05662badada9ae0d190158e03574fe69fd7cb4fc19cce856f6cee03101c7" gracePeriod=30 Oct 07 14:40:48 crc kubenswrapper[4959]: I1007 14:40:48.489065 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="78a6326b-c7db-484b-8065-481b55b8297e" containerName="manila-api-log" containerID="cri-o://be93772201bc7be59a14408abebae07a87490981f194939cf363d98caf98ce6f" gracePeriod=30 Oct 07 14:40:48 crc kubenswrapper[4959]: I1007 14:40:48.495742 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79d4ff65f9-n8c9v" event={"ID":"d46df3be-9794-4019-8099-9c3757b5b468","Type":"ContainerStarted","Data":"bc5d289fe84b985a1bf497955acff98b04cd3ccd6c6ec3676cce88e5b9fdc179"} Oct 07 14:40:48 crc kubenswrapper[4959]: I1007 14:40:48.495898 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79d4ff65f9-n8c9v" Oct 07 14:40:48 crc kubenswrapper[4959]: I1007 14:40:48.521490 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.521469326 podStartE2EDuration="3.521469326s" podCreationTimestamp="2025-10-07 14:40:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:40:48.511758 +0000 UTC m=+3330.595162325" watchObservedRunningTime="2025-10-07 14:40:48.521469326 +0000 UTC m=+3330.604873651" Oct 07 14:40:48 crc kubenswrapper[4959]: I1007 14:40:48.532897 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79d4ff65f9-n8c9v" podStartSLOduration=4.532875746 podStartE2EDuration="4.532875746s" podCreationTimestamp="2025-10-07 14:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:40:48.531438079 +0000 UTC m=+3330.614842404" watchObservedRunningTime="2025-10-07 14:40:48.532875746 +0000 UTC m=+3330.616280071" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.086912 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.233411 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78a6326b-c7db-484b-8065-481b55b8297e-logs\") pod \"78a6326b-c7db-484b-8065-481b55b8297e\" (UID: \"78a6326b-c7db-484b-8065-481b55b8297e\") " Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.233482 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78a6326b-c7db-484b-8065-481b55b8297e-etc-machine-id\") pod \"78a6326b-c7db-484b-8065-481b55b8297e\" (UID: \"78a6326b-c7db-484b-8065-481b55b8297e\") " Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.233592 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78a6326b-c7db-484b-8065-481b55b8297e-combined-ca-bundle\") pod \"78a6326b-c7db-484b-8065-481b55b8297e\" (UID: \"78a6326b-c7db-484b-8065-481b55b8297e\") " Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.233605 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/78a6326b-c7db-484b-8065-481b55b8297e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "78a6326b-c7db-484b-8065-481b55b8297e" (UID: "78a6326b-c7db-484b-8065-481b55b8297e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.233682 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsggg\" (UniqueName: \"kubernetes.io/projected/78a6326b-c7db-484b-8065-481b55b8297e-kube-api-access-jsggg\") pod \"78a6326b-c7db-484b-8065-481b55b8297e\" (UID: \"78a6326b-c7db-484b-8065-481b55b8297e\") " Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.233775 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78a6326b-c7db-484b-8065-481b55b8297e-scripts\") pod \"78a6326b-c7db-484b-8065-481b55b8297e\" (UID: \"78a6326b-c7db-484b-8065-481b55b8297e\") " Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.233804 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78a6326b-c7db-484b-8065-481b55b8297e-config-data\") pod \"78a6326b-c7db-484b-8065-481b55b8297e\" (UID: \"78a6326b-c7db-484b-8065-481b55b8297e\") " Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.233858 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78a6326b-c7db-484b-8065-481b55b8297e-config-data-custom\") pod \"78a6326b-c7db-484b-8065-481b55b8297e\" (UID: \"78a6326b-c7db-484b-8065-481b55b8297e\") " Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.233965 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78a6326b-c7db-484b-8065-481b55b8297e-logs" (OuterVolumeSpecName: "logs") pod "78a6326b-c7db-484b-8065-481b55b8297e" (UID: "78a6326b-c7db-484b-8065-481b55b8297e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.234429 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78a6326b-c7db-484b-8065-481b55b8297e-logs\") on node \"crc\" DevicePath \"\"" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.234446 4959 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78a6326b-c7db-484b-8065-481b55b8297e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.239386 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78a6326b-c7db-484b-8065-481b55b8297e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "78a6326b-c7db-484b-8065-481b55b8297e" (UID: "78a6326b-c7db-484b-8065-481b55b8297e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.243323 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78a6326b-c7db-484b-8065-481b55b8297e-kube-api-access-jsggg" (OuterVolumeSpecName: "kube-api-access-jsggg") pod "78a6326b-c7db-484b-8065-481b55b8297e" (UID: "78a6326b-c7db-484b-8065-481b55b8297e"). InnerVolumeSpecName "kube-api-access-jsggg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.262343 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78a6326b-c7db-484b-8065-481b55b8297e-scripts" (OuterVolumeSpecName: "scripts") pod "78a6326b-c7db-484b-8065-481b55b8297e" (UID: "78a6326b-c7db-484b-8065-481b55b8297e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.331024 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78a6326b-c7db-484b-8065-481b55b8297e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "78a6326b-c7db-484b-8065-481b55b8297e" (UID: "78a6326b-c7db-484b-8065-481b55b8297e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.336581 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78a6326b-c7db-484b-8065-481b55b8297e-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.336604 4959 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78a6326b-c7db-484b-8065-481b55b8297e-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.336624 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78a6326b-c7db-484b-8065-481b55b8297e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.336633 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsggg\" (UniqueName: \"kubernetes.io/projected/78a6326b-c7db-484b-8065-481b55b8297e-kube-api-access-jsggg\") on node \"crc\" DevicePath \"\"" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.354258 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78a6326b-c7db-484b-8065-481b55b8297e-config-data" (OuterVolumeSpecName: "config-data") pod "78a6326b-c7db-484b-8065-481b55b8297e" (UID: "78a6326b-c7db-484b-8065-481b55b8297e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.438242 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78a6326b-c7db-484b-8065-481b55b8297e-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.509340 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"92f09b71-3264-41e1-85b9-86ea71cffc91","Type":"ContainerStarted","Data":"258c879de3699eb7973606681429e8b77c6182a1346099c3b54ebd2732677077"} Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.509402 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"92f09b71-3264-41e1-85b9-86ea71cffc91","Type":"ContainerStarted","Data":"b1f6fb8033c51d48ecdf5784fa678abf59a9811a8fc144a2bf639c4e628e7fe8"} Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.513442 4959 generic.go:334] "Generic (PLEG): container finished" podID="78a6326b-c7db-484b-8065-481b55b8297e" containerID="6a8e05662badada9ae0d190158e03574fe69fd7cb4fc19cce856f6cee03101c7" exitCode=143 Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.513479 4959 generic.go:334] "Generic (PLEG): container finished" podID="78a6326b-c7db-484b-8065-481b55b8297e" containerID="be93772201bc7be59a14408abebae07a87490981f194939cf363d98caf98ce6f" exitCode=143 Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.513527 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.513588 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"78a6326b-c7db-484b-8065-481b55b8297e","Type":"ContainerDied","Data":"6a8e05662badada9ae0d190158e03574fe69fd7cb4fc19cce856f6cee03101c7"} Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.513619 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"78a6326b-c7db-484b-8065-481b55b8297e","Type":"ContainerDied","Data":"be93772201bc7be59a14408abebae07a87490981f194939cf363d98caf98ce6f"} Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.513630 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"78a6326b-c7db-484b-8065-481b55b8297e","Type":"ContainerDied","Data":"3d448687467142d25d309eb24914c822557f56d6acfd05eb3778c77813abaf06"} Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.513648 4959 scope.go:117] "RemoveContainer" containerID="6a8e05662badada9ae0d190158e03574fe69fd7cb4fc19cce856f6cee03101c7" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.541493 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.922842364 podStartE2EDuration="5.541471991s" podCreationTimestamp="2025-10-07 14:40:44 +0000 UTC" firstStartedPulling="2025-10-07 14:40:46.470603982 +0000 UTC m=+3328.554008307" lastFinishedPulling="2025-10-07 14:40:48.089233609 +0000 UTC m=+3330.172637934" observedRunningTime="2025-10-07 14:40:49.52767095 +0000 UTC m=+3331.611075265" watchObservedRunningTime="2025-10-07 14:40:49.541471991 +0000 UTC m=+3331.624876306" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.556896 4959 scope.go:117] "RemoveContainer" containerID="be93772201bc7be59a14408abebae07a87490981f194939cf363d98caf98ce6f" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.560424 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.585694 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-api-0"] Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.604958 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Oct 07 14:40:49 crc kubenswrapper[4959]: E1007 14:40:49.605434 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78a6326b-c7db-484b-8065-481b55b8297e" containerName="manila-api-log" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.605457 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="78a6326b-c7db-484b-8065-481b55b8297e" containerName="manila-api-log" Oct 07 14:40:49 crc kubenswrapper[4959]: E1007 14:40:49.605486 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78a6326b-c7db-484b-8065-481b55b8297e" containerName="manila-api" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.605495 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="78a6326b-c7db-484b-8065-481b55b8297e" containerName="manila-api" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.605736 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="78a6326b-c7db-484b-8065-481b55b8297e" containerName="manila-api" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.605773 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="78a6326b-c7db-484b-8065-481b55b8297e" containerName="manila-api-log" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.606901 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.607004 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.620205 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-public-svc" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.620298 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.620638 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-internal-svc" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.633772 4959 scope.go:117] "RemoveContainer" containerID="6a8e05662badada9ae0d190158e03574fe69fd7cb4fc19cce856f6cee03101c7" Oct 07 14:40:49 crc kubenswrapper[4959]: E1007 14:40:49.634350 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a8e05662badada9ae0d190158e03574fe69fd7cb4fc19cce856f6cee03101c7\": container with ID starting with 6a8e05662badada9ae0d190158e03574fe69fd7cb4fc19cce856f6cee03101c7 not found: ID does not exist" containerID="6a8e05662badada9ae0d190158e03574fe69fd7cb4fc19cce856f6cee03101c7" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.634384 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a8e05662badada9ae0d190158e03574fe69fd7cb4fc19cce856f6cee03101c7"} err="failed to get container status \"6a8e05662badada9ae0d190158e03574fe69fd7cb4fc19cce856f6cee03101c7\": rpc error: code = NotFound desc = could not find container \"6a8e05662badada9ae0d190158e03574fe69fd7cb4fc19cce856f6cee03101c7\": container with ID starting with 6a8e05662badada9ae0d190158e03574fe69fd7cb4fc19cce856f6cee03101c7 not found: ID does not exist" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.634407 4959 scope.go:117] "RemoveContainer" containerID="be93772201bc7be59a14408abebae07a87490981f194939cf363d98caf98ce6f" Oct 07 14:40:49 crc kubenswrapper[4959]: E1007 14:40:49.635658 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be93772201bc7be59a14408abebae07a87490981f194939cf363d98caf98ce6f\": container with ID starting with be93772201bc7be59a14408abebae07a87490981f194939cf363d98caf98ce6f not found: ID does not exist" containerID="be93772201bc7be59a14408abebae07a87490981f194939cf363d98caf98ce6f" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.635687 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be93772201bc7be59a14408abebae07a87490981f194939cf363d98caf98ce6f"} err="failed to get container status \"be93772201bc7be59a14408abebae07a87490981f194939cf363d98caf98ce6f\": rpc error: code = NotFound desc = could not find container \"be93772201bc7be59a14408abebae07a87490981f194939cf363d98caf98ce6f\": container with ID starting with be93772201bc7be59a14408abebae07a87490981f194939cf363d98caf98ce6f not found: ID does not exist" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.635709 4959 scope.go:117] "RemoveContainer" containerID="6a8e05662badada9ae0d190158e03574fe69fd7cb4fc19cce856f6cee03101c7" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.643599 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a8e05662badada9ae0d190158e03574fe69fd7cb4fc19cce856f6cee03101c7"} err="failed to get container status \"6a8e05662badada9ae0d190158e03574fe69fd7cb4fc19cce856f6cee03101c7\": rpc error: code = NotFound desc = could not find container \"6a8e05662badada9ae0d190158e03574fe69fd7cb4fc19cce856f6cee03101c7\": container with ID starting with 6a8e05662badada9ae0d190158e03574fe69fd7cb4fc19cce856f6cee03101c7 not found: ID does not exist" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.643661 4959 scope.go:117] "RemoveContainer" containerID="be93772201bc7be59a14408abebae07a87490981f194939cf363d98caf98ce6f" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.644850 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be93772201bc7be59a14408abebae07a87490981f194939cf363d98caf98ce6f"} err="failed to get container status \"be93772201bc7be59a14408abebae07a87490981f194939cf363d98caf98ce6f\": rpc error: code = NotFound desc = could not find container \"be93772201bc7be59a14408abebae07a87490981f194939cf363d98caf98ce6f\": container with ID starting with be93772201bc7be59a14408abebae07a87490981f194939cf363d98caf98ce6f not found: ID does not exist" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.748303 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23643eb4-c61b-4785-9a95-413794a0b756-logs\") pod \"manila-api-0\" (UID: \"23643eb4-c61b-4785-9a95-413794a0b756\") " pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.748397 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23643eb4-c61b-4785-9a95-413794a0b756-config-data-custom\") pod \"manila-api-0\" (UID: \"23643eb4-c61b-4785-9a95-413794a0b756\") " pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.748436 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/23643eb4-c61b-4785-9a95-413794a0b756-etc-machine-id\") pod \"manila-api-0\" (UID: \"23643eb4-c61b-4785-9a95-413794a0b756\") " pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.748459 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23643eb4-c61b-4785-9a95-413794a0b756-scripts\") pod \"manila-api-0\" (UID: \"23643eb4-c61b-4785-9a95-413794a0b756\") " pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.748487 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlx6g\" (UniqueName: \"kubernetes.io/projected/23643eb4-c61b-4785-9a95-413794a0b756-kube-api-access-mlx6g\") pod \"manila-api-0\" (UID: \"23643eb4-c61b-4785-9a95-413794a0b756\") " pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.748585 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/23643eb4-c61b-4785-9a95-413794a0b756-internal-tls-certs\") pod \"manila-api-0\" (UID: \"23643eb4-c61b-4785-9a95-413794a0b756\") " pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.748667 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23643eb4-c61b-4785-9a95-413794a0b756-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"23643eb4-c61b-4785-9a95-413794a0b756\") " pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.748708 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23643eb4-c61b-4785-9a95-413794a0b756-config-data\") pod \"manila-api-0\" (UID: \"23643eb4-c61b-4785-9a95-413794a0b756\") " pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.748726 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23643eb4-c61b-4785-9a95-413794a0b756-public-tls-certs\") pod \"manila-api-0\" (UID: \"23643eb4-c61b-4785-9a95-413794a0b756\") " pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.851199 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23643eb4-c61b-4785-9a95-413794a0b756-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"23643eb4-c61b-4785-9a95-413794a0b756\") " pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.851277 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23643eb4-c61b-4785-9a95-413794a0b756-config-data\") pod \"manila-api-0\" (UID: \"23643eb4-c61b-4785-9a95-413794a0b756\") " pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.851301 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23643eb4-c61b-4785-9a95-413794a0b756-public-tls-certs\") pod \"manila-api-0\" (UID: \"23643eb4-c61b-4785-9a95-413794a0b756\") " pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.851373 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23643eb4-c61b-4785-9a95-413794a0b756-logs\") pod \"manila-api-0\" (UID: \"23643eb4-c61b-4785-9a95-413794a0b756\") " pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.851409 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23643eb4-c61b-4785-9a95-413794a0b756-config-data-custom\") pod \"manila-api-0\" (UID: \"23643eb4-c61b-4785-9a95-413794a0b756\") " pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.851445 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/23643eb4-c61b-4785-9a95-413794a0b756-etc-machine-id\") pod \"manila-api-0\" (UID: \"23643eb4-c61b-4785-9a95-413794a0b756\") " pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.851468 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23643eb4-c61b-4785-9a95-413794a0b756-scripts\") pod \"manila-api-0\" (UID: \"23643eb4-c61b-4785-9a95-413794a0b756\") " pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.851492 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlx6g\" (UniqueName: \"kubernetes.io/projected/23643eb4-c61b-4785-9a95-413794a0b756-kube-api-access-mlx6g\") pod \"manila-api-0\" (UID: \"23643eb4-c61b-4785-9a95-413794a0b756\") " pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.851543 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/23643eb4-c61b-4785-9a95-413794a0b756-internal-tls-certs\") pod \"manila-api-0\" (UID: \"23643eb4-c61b-4785-9a95-413794a0b756\") " pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.851635 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/23643eb4-c61b-4785-9a95-413794a0b756-etc-machine-id\") pod \"manila-api-0\" (UID: \"23643eb4-c61b-4785-9a95-413794a0b756\") " pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.852976 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23643eb4-c61b-4785-9a95-413794a0b756-logs\") pod \"manila-api-0\" (UID: \"23643eb4-c61b-4785-9a95-413794a0b756\") " pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.858583 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23643eb4-c61b-4785-9a95-413794a0b756-scripts\") pod \"manila-api-0\" (UID: \"23643eb4-c61b-4785-9a95-413794a0b756\") " pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.859289 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23643eb4-c61b-4785-9a95-413794a0b756-config-data\") pod \"manila-api-0\" (UID: \"23643eb4-c61b-4785-9a95-413794a0b756\") " pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.860339 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/23643eb4-c61b-4785-9a95-413794a0b756-internal-tls-certs\") pod \"manila-api-0\" (UID: \"23643eb4-c61b-4785-9a95-413794a0b756\") " pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.861758 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23643eb4-c61b-4785-9a95-413794a0b756-public-tls-certs\") pod \"manila-api-0\" (UID: \"23643eb4-c61b-4785-9a95-413794a0b756\") " pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.872116 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlx6g\" (UniqueName: \"kubernetes.io/projected/23643eb4-c61b-4785-9a95-413794a0b756-kube-api-access-mlx6g\") pod \"manila-api-0\" (UID: \"23643eb4-c61b-4785-9a95-413794a0b756\") " pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.874335 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23643eb4-c61b-4785-9a95-413794a0b756-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"23643eb4-c61b-4785-9a95-413794a0b756\") " pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.889682 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23643eb4-c61b-4785-9a95-413794a0b756-config-data-custom\") pod \"manila-api-0\" (UID: \"23643eb4-c61b-4785-9a95-413794a0b756\") " pod="openstack/manila-api-0" Oct 07 14:40:49 crc kubenswrapper[4959]: I1007 14:40:49.949338 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 07 14:40:50 crc kubenswrapper[4959]: I1007 14:40:50.407738 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:40:50 crc kubenswrapper[4959]: I1007 14:40:50.408203 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bd6ea2ae-da94-4567-b894-187901295d11" containerName="ceilometer-central-agent" containerID="cri-o://4356eb32c5bc35fe66019602e00451bd08bff3a537274c3dc6a8c211974bc33d" gracePeriod=30 Oct 07 14:40:50 crc kubenswrapper[4959]: I1007 14:40:50.408752 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bd6ea2ae-da94-4567-b894-187901295d11" containerName="sg-core" containerID="cri-o://facb0655c5f056bc6a8956fc3e14daa62f6d0820180538b3f17c1e1e4a630a76" gracePeriod=30 Oct 07 14:40:50 crc kubenswrapper[4959]: I1007 14:40:50.408794 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bd6ea2ae-da94-4567-b894-187901295d11" containerName="proxy-httpd" containerID="cri-o://1484be398cc30548478f33867e65fcedcc0aa3cfb54643f6125d3cf9499fae12" gracePeriod=30 Oct 07 14:40:50 crc kubenswrapper[4959]: I1007 14:40:50.408842 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bd6ea2ae-da94-4567-b894-187901295d11" containerName="ceilometer-notification-agent" containerID="cri-o://ee10c2c91d1929fe62bea6ab20cac73239dd77669957a696ebe922a81a9c40a4" gracePeriod=30 Oct 07 14:40:50 crc kubenswrapper[4959]: I1007 14:40:50.527392 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 07 14:40:50 crc kubenswrapper[4959]: W1007 14:40:50.552693 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23643eb4_c61b_4785_9a95_413794a0b756.slice/crio-2418c77552fc28b27f8cbb311b416f0a4c1d80735f7cfe05f8fd2c0427e3619e WatchSource:0}: Error finding container 2418c77552fc28b27f8cbb311b416f0a4c1d80735f7cfe05f8fd2c0427e3619e: Status 404 returned error can't find the container with id 2418c77552fc28b27f8cbb311b416f0a4c1d80735f7cfe05f8fd2c0427e3619e Oct 07 14:40:50 crc kubenswrapper[4959]: I1007 14:40:50.680649 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78a6326b-c7db-484b-8065-481b55b8297e" path="/var/lib/kubelet/pods/78a6326b-c7db-484b-8065-481b55b8297e/volumes" Oct 07 14:40:51 crc kubenswrapper[4959]: E1007 14:40:51.182131 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd6ea2ae_da94_4567_b894_187901295d11.slice/crio-4356eb32c5bc35fe66019602e00451bd08bff3a537274c3dc6a8c211974bc33d.scope\": RecentStats: unable to find data in memory cache]" Oct 07 14:40:51 crc kubenswrapper[4959]: I1007 14:40:51.567088 4959 generic.go:334] "Generic (PLEG): container finished" podID="bd6ea2ae-da94-4567-b894-187901295d11" containerID="1484be398cc30548478f33867e65fcedcc0aa3cfb54643f6125d3cf9499fae12" exitCode=0 Oct 07 14:40:51 crc kubenswrapper[4959]: I1007 14:40:51.567554 4959 generic.go:334] "Generic (PLEG): container finished" podID="bd6ea2ae-da94-4567-b894-187901295d11" containerID="facb0655c5f056bc6a8956fc3e14daa62f6d0820180538b3f17c1e1e4a630a76" exitCode=2 Oct 07 14:40:51 crc kubenswrapper[4959]: I1007 14:40:51.567571 4959 generic.go:334] "Generic (PLEG): container finished" podID="bd6ea2ae-da94-4567-b894-187901295d11" containerID="4356eb32c5bc35fe66019602e00451bd08bff3a537274c3dc6a8c211974bc33d" exitCode=0 Oct 07 14:40:51 crc kubenswrapper[4959]: I1007 14:40:51.567142 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd6ea2ae-da94-4567-b894-187901295d11","Type":"ContainerDied","Data":"1484be398cc30548478f33867e65fcedcc0aa3cfb54643f6125d3cf9499fae12"} Oct 07 14:40:51 crc kubenswrapper[4959]: I1007 14:40:51.567662 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd6ea2ae-da94-4567-b894-187901295d11","Type":"ContainerDied","Data":"facb0655c5f056bc6a8956fc3e14daa62f6d0820180538b3f17c1e1e4a630a76"} Oct 07 14:40:51 crc kubenswrapper[4959]: I1007 14:40:51.567676 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd6ea2ae-da94-4567-b894-187901295d11","Type":"ContainerDied","Data":"4356eb32c5bc35fe66019602e00451bd08bff3a537274c3dc6a8c211974bc33d"} Oct 07 14:40:51 crc kubenswrapper[4959]: I1007 14:40:51.573847 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"23643eb4-c61b-4785-9a95-413794a0b756","Type":"ContainerStarted","Data":"2d879c7939b16c3b852222395828e2332353810e0f93b8ced96748ae38c87b00"} Oct 07 14:40:51 crc kubenswrapper[4959]: I1007 14:40:51.573905 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"23643eb4-c61b-4785-9a95-413794a0b756","Type":"ContainerStarted","Data":"2418c77552fc28b27f8cbb311b416f0a4c1d80735f7cfe05f8fd2c0427e3619e"} Oct 07 14:40:52 crc kubenswrapper[4959]: I1007 14:40:52.597175 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"23643eb4-c61b-4785-9a95-413794a0b756","Type":"ContainerStarted","Data":"ce502a3303ab6a16cbe41b45f8d9cf43b2879102edb7f2e26c7b748ca564323f"} Oct 07 14:40:52 crc kubenswrapper[4959]: I1007 14:40:52.599287 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Oct 07 14:40:52 crc kubenswrapper[4959]: I1007 14:40:52.619749 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.6197252669999997 podStartE2EDuration="3.619725267s" podCreationTimestamp="2025-10-07 14:40:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:40:52.618497006 +0000 UTC m=+3334.701901331" watchObservedRunningTime="2025-10-07 14:40:52.619725267 +0000 UTC m=+3334.703129592" Oct 07 14:40:55 crc kubenswrapper[4959]: I1007 14:40:55.110752 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Oct 07 14:40:55 crc kubenswrapper[4959]: I1007 14:40:55.320310 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79d4ff65f9-n8c9v" Oct 07 14:40:55 crc kubenswrapper[4959]: I1007 14:40:55.379195 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dbd9bf859-chpc6"] Oct 07 14:40:55 crc kubenswrapper[4959]: I1007 14:40:55.379524 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" podUID="2726035d-3257-42b4-b38a-8384e1db7d95" containerName="dnsmasq-dns" containerID="cri-o://7c8d9601815a9dd378f0655b175e064b752834b59a1de3d9f0474dbdf08e7dad" gracePeriod=10 Oct 07 14:40:55 crc kubenswrapper[4959]: I1007 14:40:55.630568 4959 generic.go:334] "Generic (PLEG): container finished" podID="2726035d-3257-42b4-b38a-8384e1db7d95" containerID="7c8d9601815a9dd378f0655b175e064b752834b59a1de3d9f0474dbdf08e7dad" exitCode=0 Oct 07 14:40:55 crc kubenswrapper[4959]: I1007 14:40:55.630618 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" event={"ID":"2726035d-3257-42b4-b38a-8384e1db7d95","Type":"ContainerDied","Data":"7c8d9601815a9dd378f0655b175e064b752834b59a1de3d9f0474dbdf08e7dad"} Oct 07 14:40:56 crc kubenswrapper[4959]: I1007 14:40:56.176025 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" Oct 07 14:40:56 crc kubenswrapper[4959]: I1007 14:40:56.342839 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-dns-svc\") pod \"2726035d-3257-42b4-b38a-8384e1db7d95\" (UID: \"2726035d-3257-42b4-b38a-8384e1db7d95\") " Oct 07 14:40:56 crc kubenswrapper[4959]: I1007 14:40:56.342908 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmpdr\" (UniqueName: \"kubernetes.io/projected/2726035d-3257-42b4-b38a-8384e1db7d95-kube-api-access-pmpdr\") pod \"2726035d-3257-42b4-b38a-8384e1db7d95\" (UID: \"2726035d-3257-42b4-b38a-8384e1db7d95\") " Oct 07 14:40:56 crc kubenswrapper[4959]: I1007 14:40:56.343046 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-config\") pod \"2726035d-3257-42b4-b38a-8384e1db7d95\" (UID: \"2726035d-3257-42b4-b38a-8384e1db7d95\") " Oct 07 14:40:56 crc kubenswrapper[4959]: I1007 14:40:56.343081 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-ovsdbserver-nb\") pod \"2726035d-3257-42b4-b38a-8384e1db7d95\" (UID: \"2726035d-3257-42b4-b38a-8384e1db7d95\") " Oct 07 14:40:56 crc kubenswrapper[4959]: I1007 14:40:56.343194 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-openstack-edpm-ipam\") pod \"2726035d-3257-42b4-b38a-8384e1db7d95\" (UID: \"2726035d-3257-42b4-b38a-8384e1db7d95\") " Oct 07 14:40:56 crc kubenswrapper[4959]: I1007 14:40:56.343267 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-ovsdbserver-sb\") pod \"2726035d-3257-42b4-b38a-8384e1db7d95\" (UID: \"2726035d-3257-42b4-b38a-8384e1db7d95\") " Oct 07 14:40:56 crc kubenswrapper[4959]: I1007 14:40:56.349214 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2726035d-3257-42b4-b38a-8384e1db7d95-kube-api-access-pmpdr" (OuterVolumeSpecName: "kube-api-access-pmpdr") pod "2726035d-3257-42b4-b38a-8384e1db7d95" (UID: "2726035d-3257-42b4-b38a-8384e1db7d95"). InnerVolumeSpecName "kube-api-access-pmpdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:40:56 crc kubenswrapper[4959]: I1007 14:40:56.399553 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-config" (OuterVolumeSpecName: "config") pod "2726035d-3257-42b4-b38a-8384e1db7d95" (UID: "2726035d-3257-42b4-b38a-8384e1db7d95"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:40:56 crc kubenswrapper[4959]: I1007 14:40:56.401946 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2726035d-3257-42b4-b38a-8384e1db7d95" (UID: "2726035d-3257-42b4-b38a-8384e1db7d95"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:40:56 crc kubenswrapper[4959]: I1007 14:40:56.403554 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2726035d-3257-42b4-b38a-8384e1db7d95" (UID: "2726035d-3257-42b4-b38a-8384e1db7d95"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:40:56 crc kubenswrapper[4959]: I1007 14:40:56.405422 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "2726035d-3257-42b4-b38a-8384e1db7d95" (UID: "2726035d-3257-42b4-b38a-8384e1db7d95"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:40:56 crc kubenswrapper[4959]: I1007 14:40:56.407035 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2726035d-3257-42b4-b38a-8384e1db7d95" (UID: "2726035d-3257-42b4-b38a-8384e1db7d95"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:40:56 crc kubenswrapper[4959]: I1007 14:40:56.446271 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-config\") on node \"crc\" DevicePath \"\"" Oct 07 14:40:56 crc kubenswrapper[4959]: I1007 14:40:56.446314 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 07 14:40:56 crc kubenswrapper[4959]: I1007 14:40:56.446341 4959 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 07 14:40:56 crc kubenswrapper[4959]: I1007 14:40:56.446357 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 07 14:40:56 crc kubenswrapper[4959]: I1007 14:40:56.446369 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2726035d-3257-42b4-b38a-8384e1db7d95-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 07 14:40:56 crc kubenswrapper[4959]: I1007 14:40:56.446380 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmpdr\" (UniqueName: \"kubernetes.io/projected/2726035d-3257-42b4-b38a-8384e1db7d95-kube-api-access-pmpdr\") on node \"crc\" DevicePath \"\"" Oct 07 14:40:56 crc kubenswrapper[4959]: I1007 14:40:56.678462 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" Oct 07 14:40:56 crc kubenswrapper[4959]: I1007 14:40:56.685436 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dbd9bf859-chpc6" event={"ID":"2726035d-3257-42b4-b38a-8384e1db7d95","Type":"ContainerDied","Data":"b98de7af5210491eaf1282c45f8431e456445aa6fafd4f71c1f30a0da388eafd"} Oct 07 14:40:56 crc kubenswrapper[4959]: I1007 14:40:56.685521 4959 scope.go:117] "RemoveContainer" containerID="7c8d9601815a9dd378f0655b175e064b752834b59a1de3d9f0474dbdf08e7dad" Oct 07 14:40:56 crc kubenswrapper[4959]: I1007 14:40:56.689594 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"418c77d5-2702-4326-93f7-738440fbe84f","Type":"ContainerStarted","Data":"e160434b947bba7491c28838c32656f376c1a00dd52824b330ed1611ea397eef"} Oct 07 14:40:56 crc kubenswrapper[4959]: I1007 14:40:56.774215 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dbd9bf859-chpc6"] Oct 07 14:40:56 crc kubenswrapper[4959]: I1007 14:40:56.775359 4959 scope.go:117] "RemoveContainer" containerID="c3f231de1ecd6f82555d9b5f33486a10f5348c3e86494bed7ff2d82de652ca65" Oct 07 14:40:56 crc kubenswrapper[4959]: I1007 14:40:56.783678 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-dbd9bf859-chpc6"] Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.443095 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.572295 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd6ea2ae-da94-4567-b894-187901295d11-log-httpd\") pod \"bd6ea2ae-da94-4567-b894-187901295d11\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.572399 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thtvv\" (UniqueName: \"kubernetes.io/projected/bd6ea2ae-da94-4567-b894-187901295d11-kube-api-access-thtvv\") pod \"bd6ea2ae-da94-4567-b894-187901295d11\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.572442 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-sg-core-conf-yaml\") pod \"bd6ea2ae-da94-4567-b894-187901295d11\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.572598 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-ceilometer-tls-certs\") pod \"bd6ea2ae-da94-4567-b894-187901295d11\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.572724 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-config-data\") pod \"bd6ea2ae-da94-4567-b894-187901295d11\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.572809 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd6ea2ae-da94-4567-b894-187901295d11-run-httpd\") pod \"bd6ea2ae-da94-4567-b894-187901295d11\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.572886 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-combined-ca-bundle\") pod \"bd6ea2ae-da94-4567-b894-187901295d11\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.572924 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-scripts\") pod \"bd6ea2ae-da94-4567-b894-187901295d11\" (UID: \"bd6ea2ae-da94-4567-b894-187901295d11\") " Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.573435 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd6ea2ae-da94-4567-b894-187901295d11-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bd6ea2ae-da94-4567-b894-187901295d11" (UID: "bd6ea2ae-da94-4567-b894-187901295d11"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.574417 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd6ea2ae-da94-4567-b894-187901295d11-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bd6ea2ae-da94-4567-b894-187901295d11" (UID: "bd6ea2ae-da94-4567-b894-187901295d11"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.593567 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-scripts" (OuterVolumeSpecName: "scripts") pod "bd6ea2ae-da94-4567-b894-187901295d11" (UID: "bd6ea2ae-da94-4567-b894-187901295d11"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.593737 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd6ea2ae-da94-4567-b894-187901295d11-kube-api-access-thtvv" (OuterVolumeSpecName: "kube-api-access-thtvv") pod "bd6ea2ae-da94-4567-b894-187901295d11" (UID: "bd6ea2ae-da94-4567-b894-187901295d11"). InnerVolumeSpecName "kube-api-access-thtvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.610198 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bd6ea2ae-da94-4567-b894-187901295d11" (UID: "bd6ea2ae-da94-4567-b894-187901295d11"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.665292 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "bd6ea2ae-da94-4567-b894-187901295d11" (UID: "bd6ea2ae-da94-4567-b894-187901295d11"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.676799 4959 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd6ea2ae-da94-4567-b894-187901295d11-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.676864 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thtvv\" (UniqueName: \"kubernetes.io/projected/bd6ea2ae-da94-4567-b894-187901295d11-kube-api-access-thtvv\") on node \"crc\" DevicePath \"\"" Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.676879 4959 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.676890 4959 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.676901 4959 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd6ea2ae-da94-4567-b894-187901295d11-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.676910 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.719452 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"418c77d5-2702-4326-93f7-738440fbe84f","Type":"ContainerStarted","Data":"31069827fdaa6ac39d0ffc9c949abf0f46ab5f520822a4156363f80d2101047e"} Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.726337 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bd6ea2ae-da94-4567-b894-187901295d11" (UID: "bd6ea2ae-da94-4567-b894-187901295d11"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.727368 4959 generic.go:334] "Generic (PLEG): container finished" podID="bd6ea2ae-da94-4567-b894-187901295d11" containerID="ee10c2c91d1929fe62bea6ab20cac73239dd77669957a696ebe922a81a9c40a4" exitCode=0 Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.727415 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd6ea2ae-da94-4567-b894-187901295d11","Type":"ContainerDied","Data":"ee10c2c91d1929fe62bea6ab20cac73239dd77669957a696ebe922a81a9c40a4"} Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.727441 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd6ea2ae-da94-4567-b894-187901295d11","Type":"ContainerDied","Data":"fbb0afb79f13cf9610b5fc3c35ae9de4ad08e1dc38695cd46f121ae59a40dc17"} Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.727459 4959 scope.go:117] "RemoveContainer" containerID="1484be398cc30548478f33867e65fcedcc0aa3cfb54643f6125d3cf9499fae12" Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.727634 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.735203 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-config-data" (OuterVolumeSpecName: "config-data") pod "bd6ea2ae-da94-4567-b894-187901295d11" (UID: "bd6ea2ae-da94-4567-b894-187901295d11"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.779621 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.780318 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd6ea2ae-da94-4567-b894-187901295d11-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.834129 4959 scope.go:117] "RemoveContainer" containerID="facb0655c5f056bc6a8956fc3e14daa62f6d0820180538b3f17c1e1e4a630a76" Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.854127 4959 scope.go:117] "RemoveContainer" containerID="ee10c2c91d1929fe62bea6ab20cac73239dd77669957a696ebe922a81a9c40a4" Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.879002 4959 scope.go:117] "RemoveContainer" containerID="4356eb32c5bc35fe66019602e00451bd08bff3a537274c3dc6a8c211974bc33d" Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.902548 4959 scope.go:117] "RemoveContainer" containerID="1484be398cc30548478f33867e65fcedcc0aa3cfb54643f6125d3cf9499fae12" Oct 07 14:40:57 crc kubenswrapper[4959]: E1007 14:40:57.903627 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1484be398cc30548478f33867e65fcedcc0aa3cfb54643f6125d3cf9499fae12\": container with ID starting with 1484be398cc30548478f33867e65fcedcc0aa3cfb54643f6125d3cf9499fae12 not found: ID does not exist" containerID="1484be398cc30548478f33867e65fcedcc0aa3cfb54643f6125d3cf9499fae12" Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.903725 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1484be398cc30548478f33867e65fcedcc0aa3cfb54643f6125d3cf9499fae12"} err="failed to get container status \"1484be398cc30548478f33867e65fcedcc0aa3cfb54643f6125d3cf9499fae12\": rpc error: code = NotFound desc = could not find container \"1484be398cc30548478f33867e65fcedcc0aa3cfb54643f6125d3cf9499fae12\": container with ID starting with 1484be398cc30548478f33867e65fcedcc0aa3cfb54643f6125d3cf9499fae12 not found: ID does not exist" Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.903793 4959 scope.go:117] "RemoveContainer" containerID="facb0655c5f056bc6a8956fc3e14daa62f6d0820180538b3f17c1e1e4a630a76" Oct 07 14:40:57 crc kubenswrapper[4959]: E1007 14:40:57.904847 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"facb0655c5f056bc6a8956fc3e14daa62f6d0820180538b3f17c1e1e4a630a76\": container with ID starting with facb0655c5f056bc6a8956fc3e14daa62f6d0820180538b3f17c1e1e4a630a76 not found: ID does not exist" containerID="facb0655c5f056bc6a8956fc3e14daa62f6d0820180538b3f17c1e1e4a630a76" Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.904899 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"facb0655c5f056bc6a8956fc3e14daa62f6d0820180538b3f17c1e1e4a630a76"} err="failed to get container status \"facb0655c5f056bc6a8956fc3e14daa62f6d0820180538b3f17c1e1e4a630a76\": rpc error: code = NotFound desc = could not find container \"facb0655c5f056bc6a8956fc3e14daa62f6d0820180538b3f17c1e1e4a630a76\": container with ID starting with facb0655c5f056bc6a8956fc3e14daa62f6d0820180538b3f17c1e1e4a630a76 not found: ID does not exist" Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.904933 4959 scope.go:117] "RemoveContainer" containerID="ee10c2c91d1929fe62bea6ab20cac73239dd77669957a696ebe922a81a9c40a4" Oct 07 14:40:57 crc kubenswrapper[4959]: E1007 14:40:57.905301 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee10c2c91d1929fe62bea6ab20cac73239dd77669957a696ebe922a81a9c40a4\": container with ID starting with ee10c2c91d1929fe62bea6ab20cac73239dd77669957a696ebe922a81a9c40a4 not found: ID does not exist" containerID="ee10c2c91d1929fe62bea6ab20cac73239dd77669957a696ebe922a81a9c40a4" Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.905332 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee10c2c91d1929fe62bea6ab20cac73239dd77669957a696ebe922a81a9c40a4"} err="failed to get container status \"ee10c2c91d1929fe62bea6ab20cac73239dd77669957a696ebe922a81a9c40a4\": rpc error: code = NotFound desc = could not find container \"ee10c2c91d1929fe62bea6ab20cac73239dd77669957a696ebe922a81a9c40a4\": container with ID starting with ee10c2c91d1929fe62bea6ab20cac73239dd77669957a696ebe922a81a9c40a4 not found: ID does not exist" Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.905347 4959 scope.go:117] "RemoveContainer" containerID="4356eb32c5bc35fe66019602e00451bd08bff3a537274c3dc6a8c211974bc33d" Oct 07 14:40:57 crc kubenswrapper[4959]: E1007 14:40:57.905711 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4356eb32c5bc35fe66019602e00451bd08bff3a537274c3dc6a8c211974bc33d\": container with ID starting with 4356eb32c5bc35fe66019602e00451bd08bff3a537274c3dc6a8c211974bc33d not found: ID does not exist" containerID="4356eb32c5bc35fe66019602e00451bd08bff3a537274c3dc6a8c211974bc33d" Oct 07 14:40:57 crc kubenswrapper[4959]: I1007 14:40:57.905767 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4356eb32c5bc35fe66019602e00451bd08bff3a537274c3dc6a8c211974bc33d"} err="failed to get container status \"4356eb32c5bc35fe66019602e00451bd08bff3a537274c3dc6a8c211974bc33d\": rpc error: code = NotFound desc = could not find container \"4356eb32c5bc35fe66019602e00451bd08bff3a537274c3dc6a8c211974bc33d\": container with ID starting with 4356eb32c5bc35fe66019602e00451bd08bff3a537274c3dc6a8c211974bc33d not found: ID does not exist" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.064295 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=4.713942094 podStartE2EDuration="14.064273237s" podCreationTimestamp="2025-10-07 14:40:44 +0000 UTC" firstStartedPulling="2025-10-07 14:40:46.654176604 +0000 UTC m=+3328.737580919" lastFinishedPulling="2025-10-07 14:40:56.004507737 +0000 UTC m=+3338.087912062" observedRunningTime="2025-10-07 14:40:57.745333147 +0000 UTC m=+3339.828737472" watchObservedRunningTime="2025-10-07 14:40:58.064273237 +0000 UTC m=+3340.147677562" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.069881 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.079848 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.112711 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:40:58 crc kubenswrapper[4959]: E1007 14:40:58.114274 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd6ea2ae-da94-4567-b894-187901295d11" containerName="sg-core" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.114304 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd6ea2ae-da94-4567-b894-187901295d11" containerName="sg-core" Oct 07 14:40:58 crc kubenswrapper[4959]: E1007 14:40:58.114330 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd6ea2ae-da94-4567-b894-187901295d11" containerName="ceilometer-central-agent" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.114340 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd6ea2ae-da94-4567-b894-187901295d11" containerName="ceilometer-central-agent" Oct 07 14:40:58 crc kubenswrapper[4959]: E1007 14:40:58.114357 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd6ea2ae-da94-4567-b894-187901295d11" containerName="ceilometer-notification-agent" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.114363 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd6ea2ae-da94-4567-b894-187901295d11" containerName="ceilometer-notification-agent" Oct 07 14:40:58 crc kubenswrapper[4959]: E1007 14:40:58.114378 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2726035d-3257-42b4-b38a-8384e1db7d95" containerName="dnsmasq-dns" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.114384 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2726035d-3257-42b4-b38a-8384e1db7d95" containerName="dnsmasq-dns" Oct 07 14:40:58 crc kubenswrapper[4959]: E1007 14:40:58.114402 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2726035d-3257-42b4-b38a-8384e1db7d95" containerName="init" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.114408 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2726035d-3257-42b4-b38a-8384e1db7d95" containerName="init" Oct 07 14:40:58 crc kubenswrapper[4959]: E1007 14:40:58.114442 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd6ea2ae-da94-4567-b894-187901295d11" containerName="proxy-httpd" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.114449 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd6ea2ae-da94-4567-b894-187901295d11" containerName="proxy-httpd" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.114672 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd6ea2ae-da94-4567-b894-187901295d11" containerName="ceilometer-central-agent" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.114692 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd6ea2ae-da94-4567-b894-187901295d11" containerName="proxy-httpd" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.114712 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd6ea2ae-da94-4567-b894-187901295d11" containerName="sg-core" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.114721 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd6ea2ae-da94-4567-b894-187901295d11" containerName="ceilometer-notification-agent" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.114732 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2726035d-3257-42b4-b38a-8384e1db7d95" containerName="dnsmasq-dns" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.116495 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.118324 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.118589 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.121597 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.126235 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.293751 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-config-data\") pod \"ceilometer-0\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " pod="openstack/ceilometer-0" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.293819 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ab758ef-5185-40e8-8eff-778e0462ff91-run-httpd\") pod \"ceilometer-0\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " pod="openstack/ceilometer-0" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.293847 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-scripts\") pod \"ceilometer-0\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " pod="openstack/ceilometer-0" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.293978 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " pod="openstack/ceilometer-0" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.294212 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ab758ef-5185-40e8-8eff-778e0462ff91-log-httpd\") pod \"ceilometer-0\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " pod="openstack/ceilometer-0" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.294480 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5jm7\" (UniqueName: \"kubernetes.io/projected/4ab758ef-5185-40e8-8eff-778e0462ff91-kube-api-access-r5jm7\") pod \"ceilometer-0\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " pod="openstack/ceilometer-0" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.294527 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " pod="openstack/ceilometer-0" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.294642 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " pod="openstack/ceilometer-0" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.395330 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-scripts\") pod \"ceilometer-0\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " pod="openstack/ceilometer-0" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.395401 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " pod="openstack/ceilometer-0" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.395446 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ab758ef-5185-40e8-8eff-778e0462ff91-log-httpd\") pod \"ceilometer-0\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " pod="openstack/ceilometer-0" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.395524 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5jm7\" (UniqueName: \"kubernetes.io/projected/4ab758ef-5185-40e8-8eff-778e0462ff91-kube-api-access-r5jm7\") pod \"ceilometer-0\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " pod="openstack/ceilometer-0" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.395547 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " pod="openstack/ceilometer-0" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.395998 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ab758ef-5185-40e8-8eff-778e0462ff91-log-httpd\") pod \"ceilometer-0\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " pod="openstack/ceilometer-0" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.396053 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " pod="openstack/ceilometer-0" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.396085 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-config-data\") pod \"ceilometer-0\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " pod="openstack/ceilometer-0" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.396520 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ab758ef-5185-40e8-8eff-778e0462ff91-run-httpd\") pod \"ceilometer-0\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " pod="openstack/ceilometer-0" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.396815 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ab758ef-5185-40e8-8eff-778e0462ff91-run-httpd\") pod \"ceilometer-0\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " pod="openstack/ceilometer-0" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.402034 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " pod="openstack/ceilometer-0" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.403269 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-scripts\") pod \"ceilometer-0\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " pod="openstack/ceilometer-0" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.416186 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-config-data\") pod \"ceilometer-0\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " pod="openstack/ceilometer-0" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.418265 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5jm7\" (UniqueName: \"kubernetes.io/projected/4ab758ef-5185-40e8-8eff-778e0462ff91-kube-api-access-r5jm7\") pod \"ceilometer-0\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " pod="openstack/ceilometer-0" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.422793 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " pod="openstack/ceilometer-0" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.432923 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " pod="openstack/ceilometer-0" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.442515 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.443611 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.673560 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2726035d-3257-42b4-b38a-8384e1db7d95" path="/var/lib/kubelet/pods/2726035d-3257-42b4-b38a-8384e1db7d95/volumes" Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.678227 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd6ea2ae-da94-4567-b894-187901295d11" path="/var/lib/kubelet/pods/bd6ea2ae-da94-4567-b894-187901295d11/volumes" Oct 07 14:40:58 crc kubenswrapper[4959]: W1007 14:40:58.948602 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ab758ef_5185_40e8_8eff_778e0462ff91.slice/crio-c332f146a53f914bb4ff509a4c8d13000e90e356e6bc514cc6644560510fd216 WatchSource:0}: Error finding container c332f146a53f914bb4ff509a4c8d13000e90e356e6bc514cc6644560510fd216: Status 404 returned error can't find the container with id c332f146a53f914bb4ff509a4c8d13000e90e356e6bc514cc6644560510fd216 Oct 07 14:40:58 crc kubenswrapper[4959]: I1007 14:40:58.961505 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:40:59 crc kubenswrapper[4959]: I1007 14:40:59.754042 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ab758ef-5185-40e8-8eff-778e0462ff91","Type":"ContainerStarted","Data":"c332f146a53f914bb4ff509a4c8d13000e90e356e6bc514cc6644560510fd216"} Oct 07 14:41:00 crc kubenswrapper[4959]: I1007 14:41:00.765594 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ab758ef-5185-40e8-8eff-778e0462ff91","Type":"ContainerStarted","Data":"8c45a17d934d94e24f94756dc25978129a7683c0f0fbdf5a12e734dd5f2acc51"} Oct 07 14:41:02 crc kubenswrapper[4959]: I1007 14:41:02.789008 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ab758ef-5185-40e8-8eff-778e0462ff91","Type":"ContainerStarted","Data":"8283b928ce7b81e8422d861cbd3f16e904c8488a3c7014df760c7c8f80e97b87"} Oct 07 14:41:04 crc kubenswrapper[4959]: I1007 14:41:04.817779 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ab758ef-5185-40e8-8eff-778e0462ff91","Type":"ContainerStarted","Data":"9e2fd5550f69430eb745ba053a274fa5ac41ab32587d105a70e0f046f6bc71de"} Oct 07 14:41:05 crc kubenswrapper[4959]: I1007 14:41:05.275742 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Oct 07 14:41:06 crc kubenswrapper[4959]: I1007 14:41:06.808276 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Oct 07 14:41:06 crc kubenswrapper[4959]: I1007 14:41:06.841763 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ab758ef-5185-40e8-8eff-778e0462ff91","Type":"ContainerStarted","Data":"31189037eebe1bea52f45902767cc389bf0f829c7d33e7ccccd5fad4e4cb79ee"} Oct 07 14:41:06 crc kubenswrapper[4959]: I1007 14:41:06.842251 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4ab758ef-5185-40e8-8eff-778e0462ff91" containerName="ceilometer-central-agent" containerID="cri-o://8c45a17d934d94e24f94756dc25978129a7683c0f0fbdf5a12e734dd5f2acc51" gracePeriod=30 Oct 07 14:41:06 crc kubenswrapper[4959]: I1007 14:41:06.842679 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4ab758ef-5185-40e8-8eff-778e0462ff91" containerName="sg-core" containerID="cri-o://9e2fd5550f69430eb745ba053a274fa5ac41ab32587d105a70e0f046f6bc71de" gracePeriod=30 Oct 07 14:41:06 crc kubenswrapper[4959]: I1007 14:41:06.842718 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 07 14:41:06 crc kubenswrapper[4959]: I1007 14:41:06.842722 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4ab758ef-5185-40e8-8eff-778e0462ff91" containerName="ceilometer-notification-agent" containerID="cri-o://8283b928ce7b81e8422d861cbd3f16e904c8488a3c7014df760c7c8f80e97b87" gracePeriod=30 Oct 07 14:41:06 crc kubenswrapper[4959]: I1007 14:41:06.842723 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4ab758ef-5185-40e8-8eff-778e0462ff91" containerName="proxy-httpd" containerID="cri-o://31189037eebe1bea52f45902767cc389bf0f829c7d33e7ccccd5fad4e4cb79ee" gracePeriod=30 Oct 07 14:41:06 crc kubenswrapper[4959]: I1007 14:41:06.877054 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Oct 07 14:41:06 crc kubenswrapper[4959]: I1007 14:41:06.877441 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="92f09b71-3264-41e1-85b9-86ea71cffc91" containerName="manila-scheduler" containerID="cri-o://b1f6fb8033c51d48ecdf5784fa678abf59a9811a8fc144a2bf639c4e628e7fe8" gracePeriod=30 Oct 07 14:41:06 crc kubenswrapper[4959]: I1007 14:41:06.877520 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="92f09b71-3264-41e1-85b9-86ea71cffc91" containerName="probe" containerID="cri-o://258c879de3699eb7973606681429e8b77c6182a1346099c3b54ebd2732677077" gracePeriod=30 Oct 07 14:41:06 crc kubenswrapper[4959]: I1007 14:41:06.891037 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.139623974 podStartE2EDuration="8.891011643s" podCreationTimestamp="2025-10-07 14:40:58 +0000 UTC" firstStartedPulling="2025-10-07 14:40:58.951315845 +0000 UTC m=+3341.034720170" lastFinishedPulling="2025-10-07 14:41:05.702703514 +0000 UTC m=+3347.786107839" observedRunningTime="2025-10-07 14:41:06.882028425 +0000 UTC m=+3348.965432770" watchObservedRunningTime="2025-10-07 14:41:06.891011643 +0000 UTC m=+3348.974415968" Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.695090 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.821002 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-combined-ca-bundle\") pod \"4ab758ef-5185-40e8-8eff-778e0462ff91\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.821190 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ab758ef-5185-40e8-8eff-778e0462ff91-run-httpd\") pod \"4ab758ef-5185-40e8-8eff-778e0462ff91\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.821228 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-sg-core-conf-yaml\") pod \"4ab758ef-5185-40e8-8eff-778e0462ff91\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.821316 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5jm7\" (UniqueName: \"kubernetes.io/projected/4ab758ef-5185-40e8-8eff-778e0462ff91-kube-api-access-r5jm7\") pod \"4ab758ef-5185-40e8-8eff-778e0462ff91\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.821375 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-scripts\") pod \"4ab758ef-5185-40e8-8eff-778e0462ff91\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.821395 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-ceilometer-tls-certs\") pod \"4ab758ef-5185-40e8-8eff-778e0462ff91\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.822413 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-config-data\") pod \"4ab758ef-5185-40e8-8eff-778e0462ff91\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.822475 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ab758ef-5185-40e8-8eff-778e0462ff91-log-httpd\") pod \"4ab758ef-5185-40e8-8eff-778e0462ff91\" (UID: \"4ab758ef-5185-40e8-8eff-778e0462ff91\") " Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.821712 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ab758ef-5185-40e8-8eff-778e0462ff91-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4ab758ef-5185-40e8-8eff-778e0462ff91" (UID: "4ab758ef-5185-40e8-8eff-778e0462ff91"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.825031 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ab758ef-5185-40e8-8eff-778e0462ff91-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4ab758ef-5185-40e8-8eff-778e0462ff91" (UID: "4ab758ef-5185-40e8-8eff-778e0462ff91"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.830441 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-scripts" (OuterVolumeSpecName: "scripts") pod "4ab758ef-5185-40e8-8eff-778e0462ff91" (UID: "4ab758ef-5185-40e8-8eff-778e0462ff91"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.830550 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ab758ef-5185-40e8-8eff-778e0462ff91-kube-api-access-r5jm7" (OuterVolumeSpecName: "kube-api-access-r5jm7") pod "4ab758ef-5185-40e8-8eff-778e0462ff91" (UID: "4ab758ef-5185-40e8-8eff-778e0462ff91"). InnerVolumeSpecName "kube-api-access-r5jm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.856479 4959 generic.go:334] "Generic (PLEG): container finished" podID="92f09b71-3264-41e1-85b9-86ea71cffc91" containerID="258c879de3699eb7973606681429e8b77c6182a1346099c3b54ebd2732677077" exitCode=0 Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.856522 4959 generic.go:334] "Generic (PLEG): container finished" podID="92f09b71-3264-41e1-85b9-86ea71cffc91" containerID="b1f6fb8033c51d48ecdf5784fa678abf59a9811a8fc144a2bf639c4e628e7fe8" exitCode=0 Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.856601 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"92f09b71-3264-41e1-85b9-86ea71cffc91","Type":"ContainerDied","Data":"258c879de3699eb7973606681429e8b77c6182a1346099c3b54ebd2732677077"} Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.856636 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"92f09b71-3264-41e1-85b9-86ea71cffc91","Type":"ContainerDied","Data":"b1f6fb8033c51d48ecdf5784fa678abf59a9811a8fc144a2bf639c4e628e7fe8"} Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.859995 4959 generic.go:334] "Generic (PLEG): container finished" podID="4ab758ef-5185-40e8-8eff-778e0462ff91" containerID="31189037eebe1bea52f45902767cc389bf0f829c7d33e7ccccd5fad4e4cb79ee" exitCode=0 Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.860019 4959 generic.go:334] "Generic (PLEG): container finished" podID="4ab758ef-5185-40e8-8eff-778e0462ff91" containerID="9e2fd5550f69430eb745ba053a274fa5ac41ab32587d105a70e0f046f6bc71de" exitCode=2 Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.860028 4959 generic.go:334] "Generic (PLEG): container finished" podID="4ab758ef-5185-40e8-8eff-778e0462ff91" containerID="8283b928ce7b81e8422d861cbd3f16e904c8488a3c7014df760c7c8f80e97b87" exitCode=0 Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.860037 4959 generic.go:334] "Generic (PLEG): container finished" podID="4ab758ef-5185-40e8-8eff-778e0462ff91" containerID="8c45a17d934d94e24f94756dc25978129a7683c0f0fbdf5a12e734dd5f2acc51" exitCode=0 Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.860056 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ab758ef-5185-40e8-8eff-778e0462ff91","Type":"ContainerDied","Data":"31189037eebe1bea52f45902767cc389bf0f829c7d33e7ccccd5fad4e4cb79ee"} Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.860076 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ab758ef-5185-40e8-8eff-778e0462ff91","Type":"ContainerDied","Data":"9e2fd5550f69430eb745ba053a274fa5ac41ab32587d105a70e0f046f6bc71de"} Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.860086 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ab758ef-5185-40e8-8eff-778e0462ff91","Type":"ContainerDied","Data":"8283b928ce7b81e8422d861cbd3f16e904c8488a3c7014df760c7c8f80e97b87"} Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.860114 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ab758ef-5185-40e8-8eff-778e0462ff91","Type":"ContainerDied","Data":"8c45a17d934d94e24f94756dc25978129a7683c0f0fbdf5a12e734dd5f2acc51"} Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.860124 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ab758ef-5185-40e8-8eff-778e0462ff91","Type":"ContainerDied","Data":"c332f146a53f914bb4ff509a4c8d13000e90e356e6bc514cc6644560510fd216"} Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.860142 4959 scope.go:117] "RemoveContainer" containerID="31189037eebe1bea52f45902767cc389bf0f829c7d33e7ccccd5fad4e4cb79ee" Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.860313 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.860363 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4ab758ef-5185-40e8-8eff-778e0462ff91" (UID: "4ab758ef-5185-40e8-8eff-778e0462ff91"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.883889 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "4ab758ef-5185-40e8-8eff-778e0462ff91" (UID: "4ab758ef-5185-40e8-8eff-778e0462ff91"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.886366 4959 scope.go:117] "RemoveContainer" containerID="9e2fd5550f69430eb745ba053a274fa5ac41ab32587d105a70e0f046f6bc71de" Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.917758 4959 scope.go:117] "RemoveContainer" containerID="8283b928ce7b81e8422d861cbd3f16e904c8488a3c7014df760c7c8f80e97b87" Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.917750 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4ab758ef-5185-40e8-8eff-778e0462ff91" (UID: "4ab758ef-5185-40e8-8eff-778e0462ff91"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.924645 4959 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ab758ef-5185-40e8-8eff-778e0462ff91-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.924677 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.924688 4959 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ab758ef-5185-40e8-8eff-778e0462ff91-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.924698 4959 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.924708 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5jm7\" (UniqueName: \"kubernetes.io/projected/4ab758ef-5185-40e8-8eff-778e0462ff91-kube-api-access-r5jm7\") on node \"crc\" DevicePath \"\"" Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.924720 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.924730 4959 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 07 14:41:07 crc kubenswrapper[4959]: I1007 14:41:07.997599 4959 scope.go:117] "RemoveContainer" containerID="8c45a17d934d94e24f94756dc25978129a7683c0f0fbdf5a12e734dd5f2acc51" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.005244 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-config-data" (OuterVolumeSpecName: "config-data") pod "4ab758ef-5185-40e8-8eff-778e0462ff91" (UID: "4ab758ef-5185-40e8-8eff-778e0462ff91"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.029512 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ab758ef-5185-40e8-8eff-778e0462ff91-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.256243 4959 scope.go:117] "RemoveContainer" containerID="31189037eebe1bea52f45902767cc389bf0f829c7d33e7ccccd5fad4e4cb79ee" Oct 07 14:41:08 crc kubenswrapper[4959]: E1007 14:41:08.256815 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31189037eebe1bea52f45902767cc389bf0f829c7d33e7ccccd5fad4e4cb79ee\": container with ID starting with 31189037eebe1bea52f45902767cc389bf0f829c7d33e7ccccd5fad4e4cb79ee not found: ID does not exist" containerID="31189037eebe1bea52f45902767cc389bf0f829c7d33e7ccccd5fad4e4cb79ee" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.256846 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31189037eebe1bea52f45902767cc389bf0f829c7d33e7ccccd5fad4e4cb79ee"} err="failed to get container status \"31189037eebe1bea52f45902767cc389bf0f829c7d33e7ccccd5fad4e4cb79ee\": rpc error: code = NotFound desc = could not find container \"31189037eebe1bea52f45902767cc389bf0f829c7d33e7ccccd5fad4e4cb79ee\": container with ID starting with 31189037eebe1bea52f45902767cc389bf0f829c7d33e7ccccd5fad4e4cb79ee not found: ID does not exist" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.256866 4959 scope.go:117] "RemoveContainer" containerID="9e2fd5550f69430eb745ba053a274fa5ac41ab32587d105a70e0f046f6bc71de" Oct 07 14:41:08 crc kubenswrapper[4959]: E1007 14:41:08.257139 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e2fd5550f69430eb745ba053a274fa5ac41ab32587d105a70e0f046f6bc71de\": container with ID starting with 9e2fd5550f69430eb745ba053a274fa5ac41ab32587d105a70e0f046f6bc71de not found: ID does not exist" containerID="9e2fd5550f69430eb745ba053a274fa5ac41ab32587d105a70e0f046f6bc71de" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.257160 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e2fd5550f69430eb745ba053a274fa5ac41ab32587d105a70e0f046f6bc71de"} err="failed to get container status \"9e2fd5550f69430eb745ba053a274fa5ac41ab32587d105a70e0f046f6bc71de\": rpc error: code = NotFound desc = could not find container \"9e2fd5550f69430eb745ba053a274fa5ac41ab32587d105a70e0f046f6bc71de\": container with ID starting with 9e2fd5550f69430eb745ba053a274fa5ac41ab32587d105a70e0f046f6bc71de not found: ID does not exist" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.257175 4959 scope.go:117] "RemoveContainer" containerID="8283b928ce7b81e8422d861cbd3f16e904c8488a3c7014df760c7c8f80e97b87" Oct 07 14:41:08 crc kubenswrapper[4959]: E1007 14:41:08.257866 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8283b928ce7b81e8422d861cbd3f16e904c8488a3c7014df760c7c8f80e97b87\": container with ID starting with 8283b928ce7b81e8422d861cbd3f16e904c8488a3c7014df760c7c8f80e97b87 not found: ID does not exist" containerID="8283b928ce7b81e8422d861cbd3f16e904c8488a3c7014df760c7c8f80e97b87" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.257887 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8283b928ce7b81e8422d861cbd3f16e904c8488a3c7014df760c7c8f80e97b87"} err="failed to get container status \"8283b928ce7b81e8422d861cbd3f16e904c8488a3c7014df760c7c8f80e97b87\": rpc error: code = NotFound desc = could not find container \"8283b928ce7b81e8422d861cbd3f16e904c8488a3c7014df760c7c8f80e97b87\": container with ID starting with 8283b928ce7b81e8422d861cbd3f16e904c8488a3c7014df760c7c8f80e97b87 not found: ID does not exist" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.257902 4959 scope.go:117] "RemoveContainer" containerID="8c45a17d934d94e24f94756dc25978129a7683c0f0fbdf5a12e734dd5f2acc51" Oct 07 14:41:08 crc kubenswrapper[4959]: E1007 14:41:08.258113 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c45a17d934d94e24f94756dc25978129a7683c0f0fbdf5a12e734dd5f2acc51\": container with ID starting with 8c45a17d934d94e24f94756dc25978129a7683c0f0fbdf5a12e734dd5f2acc51 not found: ID does not exist" containerID="8c45a17d934d94e24f94756dc25978129a7683c0f0fbdf5a12e734dd5f2acc51" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.258134 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c45a17d934d94e24f94756dc25978129a7683c0f0fbdf5a12e734dd5f2acc51"} err="failed to get container status \"8c45a17d934d94e24f94756dc25978129a7683c0f0fbdf5a12e734dd5f2acc51\": rpc error: code = NotFound desc = could not find container \"8c45a17d934d94e24f94756dc25978129a7683c0f0fbdf5a12e734dd5f2acc51\": container with ID starting with 8c45a17d934d94e24f94756dc25978129a7683c0f0fbdf5a12e734dd5f2acc51 not found: ID does not exist" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.258147 4959 scope.go:117] "RemoveContainer" containerID="31189037eebe1bea52f45902767cc389bf0f829c7d33e7ccccd5fad4e4cb79ee" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.258694 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.259064 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31189037eebe1bea52f45902767cc389bf0f829c7d33e7ccccd5fad4e4cb79ee"} err="failed to get container status \"31189037eebe1bea52f45902767cc389bf0f829c7d33e7ccccd5fad4e4cb79ee\": rpc error: code = NotFound desc = could not find container \"31189037eebe1bea52f45902767cc389bf0f829c7d33e7ccccd5fad4e4cb79ee\": container with ID starting with 31189037eebe1bea52f45902767cc389bf0f829c7d33e7ccccd5fad4e4cb79ee not found: ID does not exist" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.259079 4959 scope.go:117] "RemoveContainer" containerID="9e2fd5550f69430eb745ba053a274fa5ac41ab32587d105a70e0f046f6bc71de" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.267512 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e2fd5550f69430eb745ba053a274fa5ac41ab32587d105a70e0f046f6bc71de"} err="failed to get container status \"9e2fd5550f69430eb745ba053a274fa5ac41ab32587d105a70e0f046f6bc71de\": rpc error: code = NotFound desc = could not find container \"9e2fd5550f69430eb745ba053a274fa5ac41ab32587d105a70e0f046f6bc71de\": container with ID starting with 9e2fd5550f69430eb745ba053a274fa5ac41ab32587d105a70e0f046f6bc71de not found: ID does not exist" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.267573 4959 scope.go:117] "RemoveContainer" containerID="8283b928ce7b81e8422d861cbd3f16e904c8488a3c7014df760c7c8f80e97b87" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.269143 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8283b928ce7b81e8422d861cbd3f16e904c8488a3c7014df760c7c8f80e97b87"} err="failed to get container status \"8283b928ce7b81e8422d861cbd3f16e904c8488a3c7014df760c7c8f80e97b87\": rpc error: code = NotFound desc = could not find container \"8283b928ce7b81e8422d861cbd3f16e904c8488a3c7014df760c7c8f80e97b87\": container with ID starting with 8283b928ce7b81e8422d861cbd3f16e904c8488a3c7014df760c7c8f80e97b87 not found: ID does not exist" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.269176 4959 scope.go:117] "RemoveContainer" containerID="8c45a17d934d94e24f94756dc25978129a7683c0f0fbdf5a12e734dd5f2acc51" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.269596 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c45a17d934d94e24f94756dc25978129a7683c0f0fbdf5a12e734dd5f2acc51"} err="failed to get container status \"8c45a17d934d94e24f94756dc25978129a7683c0f0fbdf5a12e734dd5f2acc51\": rpc error: code = NotFound desc = could not find container \"8c45a17d934d94e24f94756dc25978129a7683c0f0fbdf5a12e734dd5f2acc51\": container with ID starting with 8c45a17d934d94e24f94756dc25978129a7683c0f0fbdf5a12e734dd5f2acc51 not found: ID does not exist" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.269637 4959 scope.go:117] "RemoveContainer" containerID="31189037eebe1bea52f45902767cc389bf0f829c7d33e7ccccd5fad4e4cb79ee" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.276338 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31189037eebe1bea52f45902767cc389bf0f829c7d33e7ccccd5fad4e4cb79ee"} err="failed to get container status \"31189037eebe1bea52f45902767cc389bf0f829c7d33e7ccccd5fad4e4cb79ee\": rpc error: code = NotFound desc = could not find container \"31189037eebe1bea52f45902767cc389bf0f829c7d33e7ccccd5fad4e4cb79ee\": container with ID starting with 31189037eebe1bea52f45902767cc389bf0f829c7d33e7ccccd5fad4e4cb79ee not found: ID does not exist" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.276396 4959 scope.go:117] "RemoveContainer" containerID="9e2fd5550f69430eb745ba053a274fa5ac41ab32587d105a70e0f046f6bc71de" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.277267 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e2fd5550f69430eb745ba053a274fa5ac41ab32587d105a70e0f046f6bc71de"} err="failed to get container status \"9e2fd5550f69430eb745ba053a274fa5ac41ab32587d105a70e0f046f6bc71de\": rpc error: code = NotFound desc = could not find container \"9e2fd5550f69430eb745ba053a274fa5ac41ab32587d105a70e0f046f6bc71de\": container with ID starting with 9e2fd5550f69430eb745ba053a274fa5ac41ab32587d105a70e0f046f6bc71de not found: ID does not exist" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.277352 4959 scope.go:117] "RemoveContainer" containerID="8283b928ce7b81e8422d861cbd3f16e904c8488a3c7014df760c7c8f80e97b87" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.280390 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8283b928ce7b81e8422d861cbd3f16e904c8488a3c7014df760c7c8f80e97b87"} err="failed to get container status \"8283b928ce7b81e8422d861cbd3f16e904c8488a3c7014df760c7c8f80e97b87\": rpc error: code = NotFound desc = could not find container \"8283b928ce7b81e8422d861cbd3f16e904c8488a3c7014df760c7c8f80e97b87\": container with ID starting with 8283b928ce7b81e8422d861cbd3f16e904c8488a3c7014df760c7c8f80e97b87 not found: ID does not exist" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.280415 4959 scope.go:117] "RemoveContainer" containerID="8c45a17d934d94e24f94756dc25978129a7683c0f0fbdf5a12e734dd5f2acc51" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.280973 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c45a17d934d94e24f94756dc25978129a7683c0f0fbdf5a12e734dd5f2acc51"} err="failed to get container status \"8c45a17d934d94e24f94756dc25978129a7683c0f0fbdf5a12e734dd5f2acc51\": rpc error: code = NotFound desc = could not find container \"8c45a17d934d94e24f94756dc25978129a7683c0f0fbdf5a12e734dd5f2acc51\": container with ID starting with 8c45a17d934d94e24f94756dc25978129a7683c0f0fbdf5a12e734dd5f2acc51 not found: ID does not exist" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.281020 4959 scope.go:117] "RemoveContainer" containerID="31189037eebe1bea52f45902767cc389bf0f829c7d33e7ccccd5fad4e4cb79ee" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.281432 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31189037eebe1bea52f45902767cc389bf0f829c7d33e7ccccd5fad4e4cb79ee"} err="failed to get container status \"31189037eebe1bea52f45902767cc389bf0f829c7d33e7ccccd5fad4e4cb79ee\": rpc error: code = NotFound desc = could not find container \"31189037eebe1bea52f45902767cc389bf0f829c7d33e7ccccd5fad4e4cb79ee\": container with ID starting with 31189037eebe1bea52f45902767cc389bf0f829c7d33e7ccccd5fad4e4cb79ee not found: ID does not exist" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.281459 4959 scope.go:117] "RemoveContainer" containerID="9e2fd5550f69430eb745ba053a274fa5ac41ab32587d105a70e0f046f6bc71de" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.282143 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e2fd5550f69430eb745ba053a274fa5ac41ab32587d105a70e0f046f6bc71de"} err="failed to get container status \"9e2fd5550f69430eb745ba053a274fa5ac41ab32587d105a70e0f046f6bc71de\": rpc error: code = NotFound desc = could not find container \"9e2fd5550f69430eb745ba053a274fa5ac41ab32587d105a70e0f046f6bc71de\": container with ID starting with 9e2fd5550f69430eb745ba053a274fa5ac41ab32587d105a70e0f046f6bc71de not found: ID does not exist" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.282168 4959 scope.go:117] "RemoveContainer" containerID="8283b928ce7b81e8422d861cbd3f16e904c8488a3c7014df760c7c8f80e97b87" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.282394 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8283b928ce7b81e8422d861cbd3f16e904c8488a3c7014df760c7c8f80e97b87"} err="failed to get container status \"8283b928ce7b81e8422d861cbd3f16e904c8488a3c7014df760c7c8f80e97b87\": rpc error: code = NotFound desc = could not find container \"8283b928ce7b81e8422d861cbd3f16e904c8488a3c7014df760c7c8f80e97b87\": container with ID starting with 8283b928ce7b81e8422d861cbd3f16e904c8488a3c7014df760c7c8f80e97b87 not found: ID does not exist" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.282424 4959 scope.go:117] "RemoveContainer" containerID="8c45a17d934d94e24f94756dc25978129a7683c0f0fbdf5a12e734dd5f2acc51" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.282614 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.282676 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c45a17d934d94e24f94756dc25978129a7683c0f0fbdf5a12e734dd5f2acc51"} err="failed to get container status \"8c45a17d934d94e24f94756dc25978129a7683c0f0fbdf5a12e734dd5f2acc51\": rpc error: code = NotFound desc = could not find container \"8c45a17d934d94e24f94756dc25978129a7683c0f0fbdf5a12e734dd5f2acc51\": container with ID starting with 8c45a17d934d94e24f94756dc25978129a7683c0f0fbdf5a12e734dd5f2acc51 not found: ID does not exist" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.298171 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.312177 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:41:08 crc kubenswrapper[4959]: E1007 14:41:08.312695 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92f09b71-3264-41e1-85b9-86ea71cffc91" containerName="probe" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.312716 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="92f09b71-3264-41e1-85b9-86ea71cffc91" containerName="probe" Oct 07 14:41:08 crc kubenswrapper[4959]: E1007 14:41:08.312732 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ab758ef-5185-40e8-8eff-778e0462ff91" containerName="proxy-httpd" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.312740 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ab758ef-5185-40e8-8eff-778e0462ff91" containerName="proxy-httpd" Oct 07 14:41:08 crc kubenswrapper[4959]: E1007 14:41:08.312759 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ab758ef-5185-40e8-8eff-778e0462ff91" containerName="ceilometer-notification-agent" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.312765 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ab758ef-5185-40e8-8eff-778e0462ff91" containerName="ceilometer-notification-agent" Oct 07 14:41:08 crc kubenswrapper[4959]: E1007 14:41:08.312776 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92f09b71-3264-41e1-85b9-86ea71cffc91" containerName="manila-scheduler" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.312782 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="92f09b71-3264-41e1-85b9-86ea71cffc91" containerName="manila-scheduler" Oct 07 14:41:08 crc kubenswrapper[4959]: E1007 14:41:08.312804 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ab758ef-5185-40e8-8eff-778e0462ff91" containerName="ceilometer-central-agent" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.312810 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ab758ef-5185-40e8-8eff-778e0462ff91" containerName="ceilometer-central-agent" Oct 07 14:41:08 crc kubenswrapper[4959]: E1007 14:41:08.312823 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ab758ef-5185-40e8-8eff-778e0462ff91" containerName="sg-core" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.312829 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ab758ef-5185-40e8-8eff-778e0462ff91" containerName="sg-core" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.313058 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ab758ef-5185-40e8-8eff-778e0462ff91" containerName="sg-core" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.313073 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ab758ef-5185-40e8-8eff-778e0462ff91" containerName="proxy-httpd" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.313086 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ab758ef-5185-40e8-8eff-778e0462ff91" containerName="ceilometer-notification-agent" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.313121 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="92f09b71-3264-41e1-85b9-86ea71cffc91" containerName="probe" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.313135 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="92f09b71-3264-41e1-85b9-86ea71cffc91" containerName="manila-scheduler" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.313144 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ab758ef-5185-40e8-8eff-778e0462ff91" containerName="ceilometer-central-agent" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.315148 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.321423 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.321650 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.321784 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.336679 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92f09b71-3264-41e1-85b9-86ea71cffc91-config-data\") pod \"92f09b71-3264-41e1-85b9-86ea71cffc91\" (UID: \"92f09b71-3264-41e1-85b9-86ea71cffc91\") " Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.336766 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/92f09b71-3264-41e1-85b9-86ea71cffc91-etc-machine-id\") pod \"92f09b71-3264-41e1-85b9-86ea71cffc91\" (UID: \"92f09b71-3264-41e1-85b9-86ea71cffc91\") " Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.336919 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92f09b71-3264-41e1-85b9-86ea71cffc91-scripts\") pod \"92f09b71-3264-41e1-85b9-86ea71cffc91\" (UID: \"92f09b71-3264-41e1-85b9-86ea71cffc91\") " Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.336949 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/92f09b71-3264-41e1-85b9-86ea71cffc91-config-data-custom\") pod \"92f09b71-3264-41e1-85b9-86ea71cffc91\" (UID: \"92f09b71-3264-41e1-85b9-86ea71cffc91\") " Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.337020 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftbqx\" (UniqueName: \"kubernetes.io/projected/92f09b71-3264-41e1-85b9-86ea71cffc91-kube-api-access-ftbqx\") pod \"92f09b71-3264-41e1-85b9-86ea71cffc91\" (UID: \"92f09b71-3264-41e1-85b9-86ea71cffc91\") " Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.337067 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92f09b71-3264-41e1-85b9-86ea71cffc91-combined-ca-bundle\") pod \"92f09b71-3264-41e1-85b9-86ea71cffc91\" (UID: \"92f09b71-3264-41e1-85b9-86ea71cffc91\") " Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.337979 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92f09b71-3264-41e1-85b9-86ea71cffc91-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "92f09b71-3264-41e1-85b9-86ea71cffc91" (UID: "92f09b71-3264-41e1-85b9-86ea71cffc91"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.358301 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92f09b71-3264-41e1-85b9-86ea71cffc91-scripts" (OuterVolumeSpecName: "scripts") pod "92f09b71-3264-41e1-85b9-86ea71cffc91" (UID: "92f09b71-3264-41e1-85b9-86ea71cffc91"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.358408 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92f09b71-3264-41e1-85b9-86ea71cffc91-kube-api-access-ftbqx" (OuterVolumeSpecName: "kube-api-access-ftbqx") pod "92f09b71-3264-41e1-85b9-86ea71cffc91" (UID: "92f09b71-3264-41e1-85b9-86ea71cffc91"). InnerVolumeSpecName "kube-api-access-ftbqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.369272 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92f09b71-3264-41e1-85b9-86ea71cffc91-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "92f09b71-3264-41e1-85b9-86ea71cffc91" (UID: "92f09b71-3264-41e1-85b9-86ea71cffc91"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.396295 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.405489 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92f09b71-3264-41e1-85b9-86ea71cffc91-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "92f09b71-3264-41e1-85b9-86ea71cffc91" (UID: "92f09b71-3264-41e1-85b9-86ea71cffc91"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.442734 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cc2047c-77e3-47f7-ae32-0053a6a17d25-log-httpd\") pod \"ceilometer-0\" (UID: \"0cc2047c-77e3-47f7-ae32-0053a6a17d25\") " pod="openstack/ceilometer-0" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.443138 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cc2047c-77e3-47f7-ae32-0053a6a17d25-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0cc2047c-77e3-47f7-ae32-0053a6a17d25\") " pod="openstack/ceilometer-0" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.443290 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0cc2047c-77e3-47f7-ae32-0053a6a17d25-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0cc2047c-77e3-47f7-ae32-0053a6a17d25\") " pod="openstack/ceilometer-0" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.443408 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0cc2047c-77e3-47f7-ae32-0053a6a17d25-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0cc2047c-77e3-47f7-ae32-0053a6a17d25\") " pod="openstack/ceilometer-0" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.443584 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cc2047c-77e3-47f7-ae32-0053a6a17d25-run-httpd\") pod \"ceilometer-0\" (UID: \"0cc2047c-77e3-47f7-ae32-0053a6a17d25\") " pod="openstack/ceilometer-0" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.443713 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cc2047c-77e3-47f7-ae32-0053a6a17d25-config-data\") pod \"ceilometer-0\" (UID: \"0cc2047c-77e3-47f7-ae32-0053a6a17d25\") " pod="openstack/ceilometer-0" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.443854 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6psxn\" (UniqueName: \"kubernetes.io/projected/0cc2047c-77e3-47f7-ae32-0053a6a17d25-kube-api-access-6psxn\") pod \"ceilometer-0\" (UID: \"0cc2047c-77e3-47f7-ae32-0053a6a17d25\") " pod="openstack/ceilometer-0" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.443982 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cc2047c-77e3-47f7-ae32-0053a6a17d25-scripts\") pod \"ceilometer-0\" (UID: \"0cc2047c-77e3-47f7-ae32-0053a6a17d25\") " pod="openstack/ceilometer-0" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.444116 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92f09b71-3264-41e1-85b9-86ea71cffc91-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.444209 4959 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/92f09b71-3264-41e1-85b9-86ea71cffc91-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.444328 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftbqx\" (UniqueName: \"kubernetes.io/projected/92f09b71-3264-41e1-85b9-86ea71cffc91-kube-api-access-ftbqx\") on node \"crc\" DevicePath \"\"" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.444389 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92f09b71-3264-41e1-85b9-86ea71cffc91-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.444479 4959 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/92f09b71-3264-41e1-85b9-86ea71cffc91-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.525346 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92f09b71-3264-41e1-85b9-86ea71cffc91-config-data" (OuterVolumeSpecName: "config-data") pod "92f09b71-3264-41e1-85b9-86ea71cffc91" (UID: "92f09b71-3264-41e1-85b9-86ea71cffc91"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.546257 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cc2047c-77e3-47f7-ae32-0053a6a17d25-scripts\") pod \"ceilometer-0\" (UID: \"0cc2047c-77e3-47f7-ae32-0053a6a17d25\") " pod="openstack/ceilometer-0" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.546697 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cc2047c-77e3-47f7-ae32-0053a6a17d25-log-httpd\") pod \"ceilometer-0\" (UID: \"0cc2047c-77e3-47f7-ae32-0053a6a17d25\") " pod="openstack/ceilometer-0" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.546789 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cc2047c-77e3-47f7-ae32-0053a6a17d25-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0cc2047c-77e3-47f7-ae32-0053a6a17d25\") " pod="openstack/ceilometer-0" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.546892 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0cc2047c-77e3-47f7-ae32-0053a6a17d25-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0cc2047c-77e3-47f7-ae32-0053a6a17d25\") " pod="openstack/ceilometer-0" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.547007 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0cc2047c-77e3-47f7-ae32-0053a6a17d25-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0cc2047c-77e3-47f7-ae32-0053a6a17d25\") " pod="openstack/ceilometer-0" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.547123 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cc2047c-77e3-47f7-ae32-0053a6a17d25-run-httpd\") pod \"ceilometer-0\" (UID: \"0cc2047c-77e3-47f7-ae32-0053a6a17d25\") " pod="openstack/ceilometer-0" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.547257 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cc2047c-77e3-47f7-ae32-0053a6a17d25-config-data\") pod \"ceilometer-0\" (UID: \"0cc2047c-77e3-47f7-ae32-0053a6a17d25\") " pod="openstack/ceilometer-0" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.547478 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6psxn\" (UniqueName: \"kubernetes.io/projected/0cc2047c-77e3-47f7-ae32-0053a6a17d25-kube-api-access-6psxn\") pod \"ceilometer-0\" (UID: \"0cc2047c-77e3-47f7-ae32-0053a6a17d25\") " pod="openstack/ceilometer-0" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.547653 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92f09b71-3264-41e1-85b9-86ea71cffc91-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.548734 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cc2047c-77e3-47f7-ae32-0053a6a17d25-run-httpd\") pod \"ceilometer-0\" (UID: \"0cc2047c-77e3-47f7-ae32-0053a6a17d25\") " pod="openstack/ceilometer-0" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.549607 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cc2047c-77e3-47f7-ae32-0053a6a17d25-log-httpd\") pod \"ceilometer-0\" (UID: \"0cc2047c-77e3-47f7-ae32-0053a6a17d25\") " pod="openstack/ceilometer-0" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.557378 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cc2047c-77e3-47f7-ae32-0053a6a17d25-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0cc2047c-77e3-47f7-ae32-0053a6a17d25\") " pod="openstack/ceilometer-0" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.558202 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cc2047c-77e3-47f7-ae32-0053a6a17d25-config-data\") pod \"ceilometer-0\" (UID: \"0cc2047c-77e3-47f7-ae32-0053a6a17d25\") " pod="openstack/ceilometer-0" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.558788 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cc2047c-77e3-47f7-ae32-0053a6a17d25-scripts\") pod \"ceilometer-0\" (UID: \"0cc2047c-77e3-47f7-ae32-0053a6a17d25\") " pod="openstack/ceilometer-0" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.559008 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0cc2047c-77e3-47f7-ae32-0053a6a17d25-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0cc2047c-77e3-47f7-ae32-0053a6a17d25\") " pod="openstack/ceilometer-0" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.563955 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0cc2047c-77e3-47f7-ae32-0053a6a17d25-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0cc2047c-77e3-47f7-ae32-0053a6a17d25\") " pod="openstack/ceilometer-0" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.572994 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6psxn\" (UniqueName: \"kubernetes.io/projected/0cc2047c-77e3-47f7-ae32-0053a6a17d25-kube-api-access-6psxn\") pod \"ceilometer-0\" (UID: \"0cc2047c-77e3-47f7-ae32-0053a6a17d25\") " pod="openstack/ceilometer-0" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.690088 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.690997 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ab758ef-5185-40e8-8eff-778e0462ff91" path="/var/lib/kubelet/pods/4ab758ef-5185-40e8-8eff-778e0462ff91/volumes" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.875439 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"92f09b71-3264-41e1-85b9-86ea71cffc91","Type":"ContainerDied","Data":"f39351c89b9ac5602da0f78ed2c464720a1b9d7dd07240f0a51c0f5ac9891d88"} Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.875512 4959 scope.go:117] "RemoveContainer" containerID="258c879de3699eb7973606681429e8b77c6182a1346099c3b54ebd2732677077" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.875719 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.907800 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.919502 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-scheduler-0"] Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.929901 4959 scope.go:117] "RemoveContainer" containerID="b1f6fb8033c51d48ecdf5784fa678abf59a9811a8fc144a2bf639c4e628e7fe8" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.931532 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.943025 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.945500 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Oct 07 14:41:08 crc kubenswrapper[4959]: I1007 14:41:08.946948 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 07 14:41:09 crc kubenswrapper[4959]: I1007 14:41:09.058411 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/24f33673-13e4-437a-9d3e-6528a9b81c35-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"24f33673-13e4-437a-9d3e-6528a9b81c35\") " pod="openstack/manila-scheduler-0" Oct 07 14:41:09 crc kubenswrapper[4959]: I1007 14:41:09.059423 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69hjc\" (UniqueName: \"kubernetes.io/projected/24f33673-13e4-437a-9d3e-6528a9b81c35-kube-api-access-69hjc\") pod \"manila-scheduler-0\" (UID: \"24f33673-13e4-437a-9d3e-6528a9b81c35\") " pod="openstack/manila-scheduler-0" Oct 07 14:41:09 crc kubenswrapper[4959]: I1007 14:41:09.059467 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24f33673-13e4-437a-9d3e-6528a9b81c35-scripts\") pod \"manila-scheduler-0\" (UID: \"24f33673-13e4-437a-9d3e-6528a9b81c35\") " pod="openstack/manila-scheduler-0" Oct 07 14:41:09 crc kubenswrapper[4959]: I1007 14:41:09.059498 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f33673-13e4-437a-9d3e-6528a9b81c35-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"24f33673-13e4-437a-9d3e-6528a9b81c35\") " pod="openstack/manila-scheduler-0" Oct 07 14:41:09 crc kubenswrapper[4959]: I1007 14:41:09.059732 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/24f33673-13e4-437a-9d3e-6528a9b81c35-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"24f33673-13e4-437a-9d3e-6528a9b81c35\") " pod="openstack/manila-scheduler-0" Oct 07 14:41:09 crc kubenswrapper[4959]: I1007 14:41:09.060015 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24f33673-13e4-437a-9d3e-6528a9b81c35-config-data\") pod \"manila-scheduler-0\" (UID: \"24f33673-13e4-437a-9d3e-6528a9b81c35\") " pod="openstack/manila-scheduler-0" Oct 07 14:41:09 crc kubenswrapper[4959]: I1007 14:41:09.162354 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/24f33673-13e4-437a-9d3e-6528a9b81c35-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"24f33673-13e4-437a-9d3e-6528a9b81c35\") " pod="openstack/manila-scheduler-0" Oct 07 14:41:09 crc kubenswrapper[4959]: I1007 14:41:09.162487 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24f33673-13e4-437a-9d3e-6528a9b81c35-config-data\") pod \"manila-scheduler-0\" (UID: \"24f33673-13e4-437a-9d3e-6528a9b81c35\") " pod="openstack/manila-scheduler-0" Oct 07 14:41:09 crc kubenswrapper[4959]: I1007 14:41:09.162536 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/24f33673-13e4-437a-9d3e-6528a9b81c35-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"24f33673-13e4-437a-9d3e-6528a9b81c35\") " pod="openstack/manila-scheduler-0" Oct 07 14:41:09 crc kubenswrapper[4959]: I1007 14:41:09.162672 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69hjc\" (UniqueName: \"kubernetes.io/projected/24f33673-13e4-437a-9d3e-6528a9b81c35-kube-api-access-69hjc\") pod \"manila-scheduler-0\" (UID: \"24f33673-13e4-437a-9d3e-6528a9b81c35\") " pod="openstack/manila-scheduler-0" Oct 07 14:41:09 crc kubenswrapper[4959]: I1007 14:41:09.162712 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24f33673-13e4-437a-9d3e-6528a9b81c35-scripts\") pod \"manila-scheduler-0\" (UID: \"24f33673-13e4-437a-9d3e-6528a9b81c35\") " pod="openstack/manila-scheduler-0" Oct 07 14:41:09 crc kubenswrapper[4959]: I1007 14:41:09.162830 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f33673-13e4-437a-9d3e-6528a9b81c35-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"24f33673-13e4-437a-9d3e-6528a9b81c35\") " pod="openstack/manila-scheduler-0" Oct 07 14:41:09 crc kubenswrapper[4959]: I1007 14:41:09.163686 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/24f33673-13e4-437a-9d3e-6528a9b81c35-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"24f33673-13e4-437a-9d3e-6528a9b81c35\") " pod="openstack/manila-scheduler-0" Oct 07 14:41:09 crc kubenswrapper[4959]: I1007 14:41:09.169727 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24f33673-13e4-437a-9d3e-6528a9b81c35-scripts\") pod \"manila-scheduler-0\" (UID: \"24f33673-13e4-437a-9d3e-6528a9b81c35\") " pod="openstack/manila-scheduler-0" Oct 07 14:41:09 crc kubenswrapper[4959]: I1007 14:41:09.169904 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/24f33673-13e4-437a-9d3e-6528a9b81c35-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"24f33673-13e4-437a-9d3e-6528a9b81c35\") " pod="openstack/manila-scheduler-0" Oct 07 14:41:09 crc kubenswrapper[4959]: I1007 14:41:09.170421 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f33673-13e4-437a-9d3e-6528a9b81c35-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"24f33673-13e4-437a-9d3e-6528a9b81c35\") " pod="openstack/manila-scheduler-0" Oct 07 14:41:09 crc kubenswrapper[4959]: I1007 14:41:09.171062 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24f33673-13e4-437a-9d3e-6528a9b81c35-config-data\") pod \"manila-scheduler-0\" (UID: \"24f33673-13e4-437a-9d3e-6528a9b81c35\") " pod="openstack/manila-scheduler-0" Oct 07 14:41:09 crc kubenswrapper[4959]: I1007 14:41:09.189320 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 07 14:41:09 crc kubenswrapper[4959]: I1007 14:41:09.193050 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69hjc\" (UniqueName: \"kubernetes.io/projected/24f33673-13e4-437a-9d3e-6528a9b81c35-kube-api-access-69hjc\") pod \"manila-scheduler-0\" (UID: \"24f33673-13e4-437a-9d3e-6528a9b81c35\") " pod="openstack/manila-scheduler-0" Oct 07 14:41:09 crc kubenswrapper[4959]: W1007 14:41:09.198336 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0cc2047c_77e3_47f7_ae32_0053a6a17d25.slice/crio-d892ba92545643110a071cec828536e958365a1f32c64a156f64149d24ac5a43 WatchSource:0}: Error finding container d892ba92545643110a071cec828536e958365a1f32c64a156f64149d24ac5a43: Status 404 returned error can't find the container with id d892ba92545643110a071cec828536e958365a1f32c64a156f64149d24ac5a43 Oct 07 14:41:09 crc kubenswrapper[4959]: I1007 14:41:09.266137 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 07 14:41:09 crc kubenswrapper[4959]: I1007 14:41:09.708186 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 07 14:41:09 crc kubenswrapper[4959]: I1007 14:41:09.898025 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"24f33673-13e4-437a-9d3e-6528a9b81c35","Type":"ContainerStarted","Data":"e8da85571225d7d1e0c4b2f9e1056431811b72870679f8923449368a71f39303"} Oct 07 14:41:09 crc kubenswrapper[4959]: I1007 14:41:09.900916 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cc2047c-77e3-47f7-ae32-0053a6a17d25","Type":"ContainerStarted","Data":"d892ba92545643110a071cec828536e958365a1f32c64a156f64149d24ac5a43"} Oct 07 14:41:10 crc kubenswrapper[4959]: I1007 14:41:10.685698 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92f09b71-3264-41e1-85b9-86ea71cffc91" path="/var/lib/kubelet/pods/92f09b71-3264-41e1-85b9-86ea71cffc91/volumes" Oct 07 14:41:10 crc kubenswrapper[4959]: I1007 14:41:10.911987 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cc2047c-77e3-47f7-ae32-0053a6a17d25","Type":"ContainerStarted","Data":"1935fbc3167a12d8a7a0b0c13a484454ac021d08d212bc335c706df85761b78b"} Oct 07 14:41:10 crc kubenswrapper[4959]: I1007 14:41:10.912343 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cc2047c-77e3-47f7-ae32-0053a6a17d25","Type":"ContainerStarted","Data":"41758d84d83bd0a92920601f9aa9a27aea70fc771effa9169f04665733d3c3bc"} Oct 07 14:41:10 crc kubenswrapper[4959]: I1007 14:41:10.913659 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"24f33673-13e4-437a-9d3e-6528a9b81c35","Type":"ContainerStarted","Data":"e99b27d992f7bd891adea1b221612152d2520b8fc6d7da739d2aa5ae325ce333"} Oct 07 14:41:10 crc kubenswrapper[4959]: I1007 14:41:10.913679 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"24f33673-13e4-437a-9d3e-6528a9b81c35","Type":"ContainerStarted","Data":"ebca3861d48fd66f23aafc7aed636c78f12137f81672cf2d9268de5617b47d47"} Oct 07 14:41:10 crc kubenswrapper[4959]: I1007 14:41:10.937975 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=2.93795449 podStartE2EDuration="2.93795449s" podCreationTimestamp="2025-10-07 14:41:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:41:10.932852891 +0000 UTC m=+3353.016257246" watchObservedRunningTime="2025-10-07 14:41:10.93795449 +0000 UTC m=+3353.021358815" Oct 07 14:41:11 crc kubenswrapper[4959]: I1007 14:41:11.355956 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Oct 07 14:41:11 crc kubenswrapper[4959]: I1007 14:41:11.927254 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cc2047c-77e3-47f7-ae32-0053a6a17d25","Type":"ContainerStarted","Data":"6869ae0cfd5b4d26ce35e95d0f704cc4818b8a8e81a35c34f9e3f373e141de7b"} Oct 07 14:41:12 crc kubenswrapper[4959]: I1007 14:41:12.942138 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cc2047c-77e3-47f7-ae32-0053a6a17d25","Type":"ContainerStarted","Data":"5e434246b7c17842bea58ef7a7f0990f25d374c7e22021f260da626c6f7c563f"} Oct 07 14:41:12 crc kubenswrapper[4959]: I1007 14:41:12.942630 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 07 14:41:16 crc kubenswrapper[4959]: I1007 14:41:16.850817 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Oct 07 14:41:16 crc kubenswrapper[4959]: I1007 14:41:16.881880 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=5.489272924 podStartE2EDuration="8.881861983s" podCreationTimestamp="2025-10-07 14:41:08 +0000 UTC" firstStartedPulling="2025-10-07 14:41:09.20108382 +0000 UTC m=+3351.284488135" lastFinishedPulling="2025-10-07 14:41:12.593672869 +0000 UTC m=+3354.677077194" observedRunningTime="2025-10-07 14:41:12.970890939 +0000 UTC m=+3355.054295274" watchObservedRunningTime="2025-10-07 14:41:16.881861983 +0000 UTC m=+3358.965266308" Oct 07 14:41:16 crc kubenswrapper[4959]: I1007 14:41:16.912812 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Oct 07 14:41:16 crc kubenswrapper[4959]: I1007 14:41:16.979914 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="418c77d5-2702-4326-93f7-738440fbe84f" containerName="manila-share" containerID="cri-o://e160434b947bba7491c28838c32656f376c1a00dd52824b330ed1611ea397eef" gracePeriod=30 Oct 07 14:41:16 crc kubenswrapper[4959]: I1007 14:41:16.980228 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="418c77d5-2702-4326-93f7-738440fbe84f" containerName="probe" containerID="cri-o://31069827fdaa6ac39d0ffc9c949abf0f46ab5f520822a4156363f80d2101047e" gracePeriod=30 Oct 07 14:41:17 crc kubenswrapper[4959]: I1007 14:41:17.993896 4959 generic.go:334] "Generic (PLEG): container finished" podID="418c77d5-2702-4326-93f7-738440fbe84f" containerID="31069827fdaa6ac39d0ffc9c949abf0f46ab5f520822a4156363f80d2101047e" exitCode=0 Oct 07 14:41:17 crc kubenswrapper[4959]: I1007 14:41:17.994304 4959 generic.go:334] "Generic (PLEG): container finished" podID="418c77d5-2702-4326-93f7-738440fbe84f" containerID="e160434b947bba7491c28838c32656f376c1a00dd52824b330ed1611ea397eef" exitCode=1 Oct 07 14:41:17 crc kubenswrapper[4959]: I1007 14:41:17.994088 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"418c77d5-2702-4326-93f7-738440fbe84f","Type":"ContainerDied","Data":"31069827fdaa6ac39d0ffc9c949abf0f46ab5f520822a4156363f80d2101047e"} Oct 07 14:41:17 crc kubenswrapper[4959]: I1007 14:41:17.994356 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"418c77d5-2702-4326-93f7-738440fbe84f","Type":"ContainerDied","Data":"e160434b947bba7491c28838c32656f376c1a00dd52824b330ed1611ea397eef"} Oct 07 14:41:17 crc kubenswrapper[4959]: I1007 14:41:17.994378 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"418c77d5-2702-4326-93f7-738440fbe84f","Type":"ContainerDied","Data":"5fcff236de49d79074b29fdb1bfcf60a4dc98fa107b6baea8dc4bfb2817ad0e1"} Oct 07 14:41:17 crc kubenswrapper[4959]: I1007 14:41:17.994391 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5fcff236de49d79074b29fdb1bfcf60a4dc98fa107b6baea8dc4bfb2817ad0e1" Oct 07 14:41:18 crc kubenswrapper[4959]: I1007 14:41:18.082294 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 07 14:41:18 crc kubenswrapper[4959]: I1007 14:41:18.179862 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/418c77d5-2702-4326-93f7-738440fbe84f-etc-machine-id\") pod \"418c77d5-2702-4326-93f7-738440fbe84f\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " Oct 07 14:41:18 crc kubenswrapper[4959]: I1007 14:41:18.180149 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/418c77d5-2702-4326-93f7-738440fbe84f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "418c77d5-2702-4326-93f7-738440fbe84f" (UID: "418c77d5-2702-4326-93f7-738440fbe84f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 14:41:18 crc kubenswrapper[4959]: I1007 14:41:18.180533 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqh7w\" (UniqueName: \"kubernetes.io/projected/418c77d5-2702-4326-93f7-738440fbe84f-kube-api-access-bqh7w\") pod \"418c77d5-2702-4326-93f7-738440fbe84f\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " Oct 07 14:41:18 crc kubenswrapper[4959]: I1007 14:41:18.180667 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/418c77d5-2702-4326-93f7-738440fbe84f-combined-ca-bundle\") pod \"418c77d5-2702-4326-93f7-738440fbe84f\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " Oct 07 14:41:18 crc kubenswrapper[4959]: I1007 14:41:18.180697 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/418c77d5-2702-4326-93f7-738440fbe84f-ceph\") pod \"418c77d5-2702-4326-93f7-738440fbe84f\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " Oct 07 14:41:18 crc kubenswrapper[4959]: I1007 14:41:18.180743 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/418c77d5-2702-4326-93f7-738440fbe84f-config-data\") pod \"418c77d5-2702-4326-93f7-738440fbe84f\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " Oct 07 14:41:18 crc kubenswrapper[4959]: I1007 14:41:18.180772 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/418c77d5-2702-4326-93f7-738440fbe84f-config-data-custom\") pod \"418c77d5-2702-4326-93f7-738440fbe84f\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " Oct 07 14:41:18 crc kubenswrapper[4959]: I1007 14:41:18.180803 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/418c77d5-2702-4326-93f7-738440fbe84f-scripts\") pod \"418c77d5-2702-4326-93f7-738440fbe84f\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " Oct 07 14:41:18 crc kubenswrapper[4959]: I1007 14:41:18.180828 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/418c77d5-2702-4326-93f7-738440fbe84f-var-lib-manila\") pod \"418c77d5-2702-4326-93f7-738440fbe84f\" (UID: \"418c77d5-2702-4326-93f7-738440fbe84f\") " Oct 07 14:41:18 crc kubenswrapper[4959]: I1007 14:41:18.181261 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/418c77d5-2702-4326-93f7-738440fbe84f-var-lib-manila" (OuterVolumeSpecName: "var-lib-manila") pod "418c77d5-2702-4326-93f7-738440fbe84f" (UID: "418c77d5-2702-4326-93f7-738440fbe84f"). InnerVolumeSpecName "var-lib-manila". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 14:41:18 crc kubenswrapper[4959]: I1007 14:41:18.181278 4959 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/418c77d5-2702-4326-93f7-738440fbe84f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 07 14:41:18 crc kubenswrapper[4959]: I1007 14:41:18.188347 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/418c77d5-2702-4326-93f7-738440fbe84f-kube-api-access-bqh7w" (OuterVolumeSpecName: "kube-api-access-bqh7w") pod "418c77d5-2702-4326-93f7-738440fbe84f" (UID: "418c77d5-2702-4326-93f7-738440fbe84f"). InnerVolumeSpecName "kube-api-access-bqh7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:41:18 crc kubenswrapper[4959]: I1007 14:41:18.188468 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/418c77d5-2702-4326-93f7-738440fbe84f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "418c77d5-2702-4326-93f7-738440fbe84f" (UID: "418c77d5-2702-4326-93f7-738440fbe84f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:41:18 crc kubenswrapper[4959]: I1007 14:41:18.191084 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/418c77d5-2702-4326-93f7-738440fbe84f-ceph" (OuterVolumeSpecName: "ceph") pod "418c77d5-2702-4326-93f7-738440fbe84f" (UID: "418c77d5-2702-4326-93f7-738440fbe84f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:41:18 crc kubenswrapper[4959]: I1007 14:41:18.200328 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/418c77d5-2702-4326-93f7-738440fbe84f-scripts" (OuterVolumeSpecName: "scripts") pod "418c77d5-2702-4326-93f7-738440fbe84f" (UID: "418c77d5-2702-4326-93f7-738440fbe84f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:41:18 crc kubenswrapper[4959]: I1007 14:41:18.269177 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/418c77d5-2702-4326-93f7-738440fbe84f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "418c77d5-2702-4326-93f7-738440fbe84f" (UID: "418c77d5-2702-4326-93f7-738440fbe84f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:41:18 crc kubenswrapper[4959]: I1007 14:41:18.283767 4959 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/418c77d5-2702-4326-93f7-738440fbe84f-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 07 14:41:18 crc kubenswrapper[4959]: I1007 14:41:18.283896 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/418c77d5-2702-4326-93f7-738440fbe84f-scripts\") on node \"crc\" DevicePath \"\"" Oct 07 14:41:18 crc kubenswrapper[4959]: I1007 14:41:18.283967 4959 reconciler_common.go:293] "Volume detached for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/418c77d5-2702-4326-93f7-738440fbe84f-var-lib-manila\") on node \"crc\" DevicePath \"\"" Oct 07 14:41:18 crc kubenswrapper[4959]: I1007 14:41:18.284034 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqh7w\" (UniqueName: \"kubernetes.io/projected/418c77d5-2702-4326-93f7-738440fbe84f-kube-api-access-bqh7w\") on node \"crc\" DevicePath \"\"" Oct 07 14:41:18 crc kubenswrapper[4959]: I1007 14:41:18.284122 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/418c77d5-2702-4326-93f7-738440fbe84f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 14:41:18 crc kubenswrapper[4959]: I1007 14:41:18.284205 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/418c77d5-2702-4326-93f7-738440fbe84f-ceph\") on node \"crc\" DevicePath \"\"" Oct 07 14:41:18 crc kubenswrapper[4959]: I1007 14:41:18.312475 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/418c77d5-2702-4326-93f7-738440fbe84f-config-data" (OuterVolumeSpecName: "config-data") pod "418c77d5-2702-4326-93f7-738440fbe84f" (UID: "418c77d5-2702-4326-93f7-738440fbe84f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:41:18 crc kubenswrapper[4959]: I1007 14:41:18.386074 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/418c77d5-2702-4326-93f7-738440fbe84f-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.000666 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.031525 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.043443 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-share-share1-0"] Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.066352 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Oct 07 14:41:19 crc kubenswrapper[4959]: E1007 14:41:19.066900 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="418c77d5-2702-4326-93f7-738440fbe84f" containerName="manila-share" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.066920 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="418c77d5-2702-4326-93f7-738440fbe84f" containerName="manila-share" Oct 07 14:41:19 crc kubenswrapper[4959]: E1007 14:41:19.066953 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="418c77d5-2702-4326-93f7-738440fbe84f" containerName="probe" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.066959 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="418c77d5-2702-4326-93f7-738440fbe84f" containerName="probe" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.067185 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="418c77d5-2702-4326-93f7-738440fbe84f" containerName="probe" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.067212 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="418c77d5-2702-4326-93f7-738440fbe84f" containerName="manila-share" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.068556 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.071482 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.082975 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.107602 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/59f4d3f2-ccf2-4724-aa24-2de193d2a2bc-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"59f4d3f2-ccf2-4724-aa24-2de193d2a2bc\") " pod="openstack/manila-share-share1-0" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.107741 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59f4d3f2-ccf2-4724-aa24-2de193d2a2bc-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"59f4d3f2-ccf2-4724-aa24-2de193d2a2bc\") " pod="openstack/manila-share-share1-0" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.107873 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/59f4d3f2-ccf2-4724-aa24-2de193d2a2bc-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"59f4d3f2-ccf2-4724-aa24-2de193d2a2bc\") " pod="openstack/manila-share-share1-0" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.107961 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59f4d3f2-ccf2-4724-aa24-2de193d2a2bc-scripts\") pod \"manila-share-share1-0\" (UID: \"59f4d3f2-ccf2-4724-aa24-2de193d2a2bc\") " pod="openstack/manila-share-share1-0" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.107998 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/59f4d3f2-ccf2-4724-aa24-2de193d2a2bc-ceph\") pod \"manila-share-share1-0\" (UID: \"59f4d3f2-ccf2-4724-aa24-2de193d2a2bc\") " pod="openstack/manila-share-share1-0" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.108409 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59f4d3f2-ccf2-4724-aa24-2de193d2a2bc-config-data\") pod \"manila-share-share1-0\" (UID: \"59f4d3f2-ccf2-4724-aa24-2de193d2a2bc\") " pod="openstack/manila-share-share1-0" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.108472 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wp7p\" (UniqueName: \"kubernetes.io/projected/59f4d3f2-ccf2-4724-aa24-2de193d2a2bc-kube-api-access-8wp7p\") pod \"manila-share-share1-0\" (UID: \"59f4d3f2-ccf2-4724-aa24-2de193d2a2bc\") " pod="openstack/manila-share-share1-0" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.108516 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/59f4d3f2-ccf2-4724-aa24-2de193d2a2bc-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"59f4d3f2-ccf2-4724-aa24-2de193d2a2bc\") " pod="openstack/manila-share-share1-0" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.211302 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/59f4d3f2-ccf2-4724-aa24-2de193d2a2bc-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"59f4d3f2-ccf2-4724-aa24-2de193d2a2bc\") " pod="openstack/manila-share-share1-0" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.211413 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59f4d3f2-ccf2-4724-aa24-2de193d2a2bc-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"59f4d3f2-ccf2-4724-aa24-2de193d2a2bc\") " pod="openstack/manila-share-share1-0" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.211442 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/59f4d3f2-ccf2-4724-aa24-2de193d2a2bc-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"59f4d3f2-ccf2-4724-aa24-2de193d2a2bc\") " pod="openstack/manila-share-share1-0" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.211494 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59f4d3f2-ccf2-4724-aa24-2de193d2a2bc-scripts\") pod \"manila-share-share1-0\" (UID: \"59f4d3f2-ccf2-4724-aa24-2de193d2a2bc\") " pod="openstack/manila-share-share1-0" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.211517 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/59f4d3f2-ccf2-4724-aa24-2de193d2a2bc-ceph\") pod \"manila-share-share1-0\" (UID: \"59f4d3f2-ccf2-4724-aa24-2de193d2a2bc\") " pod="openstack/manila-share-share1-0" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.211512 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/59f4d3f2-ccf2-4724-aa24-2de193d2a2bc-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"59f4d3f2-ccf2-4724-aa24-2de193d2a2bc\") " pod="openstack/manila-share-share1-0" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.211558 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59f4d3f2-ccf2-4724-aa24-2de193d2a2bc-config-data\") pod \"manila-share-share1-0\" (UID: \"59f4d3f2-ccf2-4724-aa24-2de193d2a2bc\") " pod="openstack/manila-share-share1-0" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.211632 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wp7p\" (UniqueName: \"kubernetes.io/projected/59f4d3f2-ccf2-4724-aa24-2de193d2a2bc-kube-api-access-8wp7p\") pod \"manila-share-share1-0\" (UID: \"59f4d3f2-ccf2-4724-aa24-2de193d2a2bc\") " pod="openstack/manila-share-share1-0" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.211652 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/59f4d3f2-ccf2-4724-aa24-2de193d2a2bc-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"59f4d3f2-ccf2-4724-aa24-2de193d2a2bc\") " pod="openstack/manila-share-share1-0" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.211677 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/59f4d3f2-ccf2-4724-aa24-2de193d2a2bc-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"59f4d3f2-ccf2-4724-aa24-2de193d2a2bc\") " pod="openstack/manila-share-share1-0" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.219739 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/59f4d3f2-ccf2-4724-aa24-2de193d2a2bc-ceph\") pod \"manila-share-share1-0\" (UID: \"59f4d3f2-ccf2-4724-aa24-2de193d2a2bc\") " pod="openstack/manila-share-share1-0" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.220611 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59f4d3f2-ccf2-4724-aa24-2de193d2a2bc-scripts\") pod \"manila-share-share1-0\" (UID: \"59f4d3f2-ccf2-4724-aa24-2de193d2a2bc\") " pod="openstack/manila-share-share1-0" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.220940 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/59f4d3f2-ccf2-4724-aa24-2de193d2a2bc-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"59f4d3f2-ccf2-4724-aa24-2de193d2a2bc\") " pod="openstack/manila-share-share1-0" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.227086 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59f4d3f2-ccf2-4724-aa24-2de193d2a2bc-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"59f4d3f2-ccf2-4724-aa24-2de193d2a2bc\") " pod="openstack/manila-share-share1-0" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.232170 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59f4d3f2-ccf2-4724-aa24-2de193d2a2bc-config-data\") pod \"manila-share-share1-0\" (UID: \"59f4d3f2-ccf2-4724-aa24-2de193d2a2bc\") " pod="openstack/manila-share-share1-0" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.236751 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wp7p\" (UniqueName: \"kubernetes.io/projected/59f4d3f2-ccf2-4724-aa24-2de193d2a2bc-kube-api-access-8wp7p\") pod \"manila-share-share1-0\" (UID: \"59f4d3f2-ccf2-4724-aa24-2de193d2a2bc\") " pod="openstack/manila-share-share1-0" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.266453 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Oct 07 14:41:19 crc kubenswrapper[4959]: I1007 14:41:19.391607 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 07 14:41:20 crc kubenswrapper[4959]: I1007 14:41:20.005646 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 07 14:41:20 crc kubenswrapper[4959]: I1007 14:41:20.630771 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:41:20 crc kubenswrapper[4959]: I1007 14:41:20.631908 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:41:20 crc kubenswrapper[4959]: I1007 14:41:20.670913 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="418c77d5-2702-4326-93f7-738440fbe84f" path="/var/lib/kubelet/pods/418c77d5-2702-4326-93f7-738440fbe84f/volumes" Oct 07 14:41:21 crc kubenswrapper[4959]: I1007 14:41:21.021142 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"59f4d3f2-ccf2-4724-aa24-2de193d2a2bc","Type":"ContainerStarted","Data":"2d46914e963c9fe1cdebec11b70ef526707c8fe7fe25d47f38414cc41ce04dfa"} Oct 07 14:41:21 crc kubenswrapper[4959]: I1007 14:41:21.021198 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"59f4d3f2-ccf2-4724-aa24-2de193d2a2bc","Type":"ContainerStarted","Data":"f95c34507be8b1d528479ac0a6c9d2c9805db03567e98ac8c84ebaeb18ca676e"} Oct 07 14:41:21 crc kubenswrapper[4959]: I1007 14:41:21.021210 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"59f4d3f2-ccf2-4724-aa24-2de193d2a2bc","Type":"ContainerStarted","Data":"c6fafd101d5c3f402562e6ed15e77545b42178c424f9529f523fe4e2fc1942db"} Oct 07 14:41:21 crc kubenswrapper[4959]: I1007 14:41:21.046055 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=2.046013657 podStartE2EDuration="2.046013657s" podCreationTimestamp="2025-10-07 14:41:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:41:21.040698042 +0000 UTC m=+3363.124102377" watchObservedRunningTime="2025-10-07 14:41:21.046013657 +0000 UTC m=+3363.129417982" Oct 07 14:41:22 crc kubenswrapper[4959]: E1007 14:41:22.059576 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod418c77d5_2702_4326_93f7_738440fbe84f.slice\": RecentStats: unable to find data in memory cache]" Oct 07 14:41:29 crc kubenswrapper[4959]: I1007 14:41:29.392753 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Oct 07 14:41:30 crc kubenswrapper[4959]: I1007 14:41:30.913969 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Oct 07 14:41:32 crc kubenswrapper[4959]: E1007 14:41:32.331948 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod418c77d5_2702_4326_93f7_738440fbe84f.slice\": RecentStats: unable to find data in memory cache]" Oct 07 14:41:38 crc kubenswrapper[4959]: I1007 14:41:38.700490 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 07 14:41:41 crc kubenswrapper[4959]: I1007 14:41:41.121441 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Oct 07 14:41:42 crc kubenswrapper[4959]: E1007 14:41:42.592596 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod418c77d5_2702_4326_93f7_738440fbe84f.slice\": RecentStats: unable to find data in memory cache]" Oct 07 14:41:50 crc kubenswrapper[4959]: I1007 14:41:50.629757 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:41:50 crc kubenswrapper[4959]: I1007 14:41:50.630450 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:41:52 crc kubenswrapper[4959]: E1007 14:41:52.897559 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod418c77d5_2702_4326_93f7_738440fbe84f.slice\": RecentStats: unable to find data in memory cache]" Oct 07 14:42:03 crc kubenswrapper[4959]: E1007 14:42:03.168083 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod418c77d5_2702_4326_93f7_738440fbe84f.slice\": RecentStats: unable to find data in memory cache]" Oct 07 14:42:13 crc kubenswrapper[4959]: E1007 14:42:13.421192 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod418c77d5_2702_4326_93f7_738440fbe84f.slice\": RecentStats: unable to find data in memory cache]" Oct 07 14:42:20 crc kubenswrapper[4959]: I1007 14:42:20.630365 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:42:20 crc kubenswrapper[4959]: I1007 14:42:20.631980 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:42:20 crc kubenswrapper[4959]: I1007 14:42:20.632133 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 14:42:20 crc kubenswrapper[4959]: I1007 14:42:20.634358 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d95ff1c6b77cf13243ed8e843bbaced75a795144cd83d3ae4f932c4a56aa2beb"} pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 14:42:20 crc kubenswrapper[4959]: I1007 14:42:20.634501 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" containerID="cri-o://d95ff1c6b77cf13243ed8e843bbaced75a795144cd83d3ae4f932c4a56aa2beb" gracePeriod=600 Oct 07 14:42:21 crc kubenswrapper[4959]: I1007 14:42:21.713332 4959 generic.go:334] "Generic (PLEG): container finished" podID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerID="d95ff1c6b77cf13243ed8e843bbaced75a795144cd83d3ae4f932c4a56aa2beb" exitCode=0 Oct 07 14:42:21 crc kubenswrapper[4959]: I1007 14:42:21.713461 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerDied","Data":"d95ff1c6b77cf13243ed8e843bbaced75a795144cd83d3ae4f932c4a56aa2beb"} Oct 07 14:42:21 crc kubenswrapper[4959]: I1007 14:42:21.714291 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerStarted","Data":"33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f"} Oct 07 14:42:21 crc kubenswrapper[4959]: I1007 14:42:21.714325 4959 scope.go:117] "RemoveContainer" containerID="5c727a08b106296927c9f8536b3664c726df30b6820379e308467063dcbc518d" Oct 07 14:42:39 crc kubenswrapper[4959]: I1007 14:42:39.958970 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5d4b45574b-njslg"] Oct 07 14:42:39 crc kubenswrapper[4959]: I1007 14:42:39.961365 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-5d4b45574b-njslg" Oct 07 14:42:39 crc kubenswrapper[4959]: I1007 14:42:39.987320 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5d4b45574b-njslg"] Oct 07 14:42:40 crc kubenswrapper[4959]: I1007 14:42:40.066645 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtjj8\" (UniqueName: \"kubernetes.io/projected/92db920f-830b-4d30-88c6-e9ebb8b83d08-kube-api-access-jtjj8\") pod \"openstack-operator-controller-operator-5d4b45574b-njslg\" (UID: \"92db920f-830b-4d30-88c6-e9ebb8b83d08\") " pod="openstack-operators/openstack-operator-controller-operator-5d4b45574b-njslg" Oct 07 14:42:40 crc kubenswrapper[4959]: I1007 14:42:40.169224 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtjj8\" (UniqueName: \"kubernetes.io/projected/92db920f-830b-4d30-88c6-e9ebb8b83d08-kube-api-access-jtjj8\") pod \"openstack-operator-controller-operator-5d4b45574b-njslg\" (UID: \"92db920f-830b-4d30-88c6-e9ebb8b83d08\") " pod="openstack-operators/openstack-operator-controller-operator-5d4b45574b-njslg" Oct 07 14:42:40 crc kubenswrapper[4959]: I1007 14:42:40.200090 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtjj8\" (UniqueName: \"kubernetes.io/projected/92db920f-830b-4d30-88c6-e9ebb8b83d08-kube-api-access-jtjj8\") pod \"openstack-operator-controller-operator-5d4b45574b-njslg\" (UID: \"92db920f-830b-4d30-88c6-e9ebb8b83d08\") " pod="openstack-operators/openstack-operator-controller-operator-5d4b45574b-njslg" Oct 07 14:42:40 crc kubenswrapper[4959]: I1007 14:42:40.287450 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-5d4b45574b-njslg" Oct 07 14:42:40 crc kubenswrapper[4959]: I1007 14:42:40.862330 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5d4b45574b-njslg"] Oct 07 14:42:40 crc kubenswrapper[4959]: W1007 14:42:40.867930 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92db920f_830b_4d30_88c6_e9ebb8b83d08.slice/crio-1efef782ddab2d5e91b1744e80a55deea2ceb6afdcfb68abdbff41527422d298 WatchSource:0}: Error finding container 1efef782ddab2d5e91b1744e80a55deea2ceb6afdcfb68abdbff41527422d298: Status 404 returned error can't find the container with id 1efef782ddab2d5e91b1744e80a55deea2ceb6afdcfb68abdbff41527422d298 Oct 07 14:42:40 crc kubenswrapper[4959]: I1007 14:42:40.947163 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5d4b45574b-njslg" event={"ID":"92db920f-830b-4d30-88c6-e9ebb8b83d08","Type":"ContainerStarted","Data":"1efef782ddab2d5e91b1744e80a55deea2ceb6afdcfb68abdbff41527422d298"} Oct 07 14:42:41 crc kubenswrapper[4959]: I1007 14:42:41.958681 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5d4b45574b-njslg" event={"ID":"92db920f-830b-4d30-88c6-e9ebb8b83d08","Type":"ContainerStarted","Data":"84bc4903e9c7ace0ed8fe3b35a9f8b6ba4494f26a56fa16b3e19f2ee16452b79"} Oct 07 14:42:41 crc kubenswrapper[4959]: I1007 14:42:41.959328 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5d4b45574b-njslg" event={"ID":"92db920f-830b-4d30-88c6-e9ebb8b83d08","Type":"ContainerStarted","Data":"63b83115d52403546bb32e5d5aff495b6df130f687fa0eee38805fef9dc4c16f"} Oct 07 14:42:41 crc kubenswrapper[4959]: I1007 14:42:41.959369 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-5d4b45574b-njslg" Oct 07 14:42:41 crc kubenswrapper[4959]: I1007 14:42:41.993610 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-5d4b45574b-njslg" podStartSLOduration=2.9935882400000002 podStartE2EDuration="2.99358824s" podCreationTimestamp="2025-10-07 14:42:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 14:42:41.988349417 +0000 UTC m=+3444.071753742" watchObservedRunningTime="2025-10-07 14:42:41.99358824 +0000 UTC m=+3444.076992565" Oct 07 14:42:50 crc kubenswrapper[4959]: I1007 14:42:50.291436 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-5d4b45574b-njslg" Oct 07 14:42:50 crc kubenswrapper[4959]: I1007 14:42:50.397831 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-57bc4467bb-cdctp"] Oct 07 14:42:50 crc kubenswrapper[4959]: I1007 14:42:50.398076 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-controller-operator-57bc4467bb-cdctp" podUID="34a5b41b-5e32-42e7-a00a-9942fd069d2b" containerName="operator" containerID="cri-o://3eadc7632c0f224744bee6c3e076dcd53e1d7cbfa2f4c07b399d1d4c8a4cdada" gracePeriod=10 Oct 07 14:42:50 crc kubenswrapper[4959]: I1007 14:42:50.398171 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-controller-operator-57bc4467bb-cdctp" podUID="34a5b41b-5e32-42e7-a00a-9942fd069d2b" containerName="kube-rbac-proxy" containerID="cri-o://5918eddd4d8d7f881902acc813928866b3fa0687a6f0763098b27756f1317ad8" gracePeriod=10 Oct 07 14:42:50 crc kubenswrapper[4959]: I1007 14:42:50.962143 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-57bc4467bb-cdctp" Oct 07 14:42:51 crc kubenswrapper[4959]: I1007 14:42:51.063593 4959 generic.go:334] "Generic (PLEG): container finished" podID="34a5b41b-5e32-42e7-a00a-9942fd069d2b" containerID="5918eddd4d8d7f881902acc813928866b3fa0687a6f0763098b27756f1317ad8" exitCode=0 Oct 07 14:42:51 crc kubenswrapper[4959]: I1007 14:42:51.063667 4959 generic.go:334] "Generic (PLEG): container finished" podID="34a5b41b-5e32-42e7-a00a-9942fd069d2b" containerID="3eadc7632c0f224744bee6c3e076dcd53e1d7cbfa2f4c07b399d1d4c8a4cdada" exitCode=0 Oct 07 14:42:51 crc kubenswrapper[4959]: I1007 14:42:51.063698 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-57bc4467bb-cdctp" event={"ID":"34a5b41b-5e32-42e7-a00a-9942fd069d2b","Type":"ContainerDied","Data":"5918eddd4d8d7f881902acc813928866b3fa0687a6f0763098b27756f1317ad8"} Oct 07 14:42:51 crc kubenswrapper[4959]: I1007 14:42:51.063763 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-57bc4467bb-cdctp" event={"ID":"34a5b41b-5e32-42e7-a00a-9942fd069d2b","Type":"ContainerDied","Data":"3eadc7632c0f224744bee6c3e076dcd53e1d7cbfa2f4c07b399d1d4c8a4cdada"} Oct 07 14:42:51 crc kubenswrapper[4959]: I1007 14:42:51.063780 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-57bc4467bb-cdctp" event={"ID":"34a5b41b-5e32-42e7-a00a-9942fd069d2b","Type":"ContainerDied","Data":"cfce20a2959e727db7ec7895039c502645d4eb6e64bb532815dae078de4a448a"} Oct 07 14:42:51 crc kubenswrapper[4959]: I1007 14:42:51.063798 4959 scope.go:117] "RemoveContainer" containerID="5918eddd4d8d7f881902acc813928866b3fa0687a6f0763098b27756f1317ad8" Oct 07 14:42:51 crc kubenswrapper[4959]: I1007 14:42:51.064078 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-57bc4467bb-cdctp" Oct 07 14:42:51 crc kubenswrapper[4959]: I1007 14:42:51.087851 4959 scope.go:117] "RemoveContainer" containerID="3eadc7632c0f224744bee6c3e076dcd53e1d7cbfa2f4c07b399d1d4c8a4cdada" Oct 07 14:42:51 crc kubenswrapper[4959]: I1007 14:42:51.146230 4959 scope.go:117] "RemoveContainer" containerID="5918eddd4d8d7f881902acc813928866b3fa0687a6f0763098b27756f1317ad8" Oct 07 14:42:51 crc kubenswrapper[4959]: E1007 14:42:51.146846 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5918eddd4d8d7f881902acc813928866b3fa0687a6f0763098b27756f1317ad8\": container with ID starting with 5918eddd4d8d7f881902acc813928866b3fa0687a6f0763098b27756f1317ad8 not found: ID does not exist" containerID="5918eddd4d8d7f881902acc813928866b3fa0687a6f0763098b27756f1317ad8" Oct 07 14:42:51 crc kubenswrapper[4959]: I1007 14:42:51.146877 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5918eddd4d8d7f881902acc813928866b3fa0687a6f0763098b27756f1317ad8"} err="failed to get container status \"5918eddd4d8d7f881902acc813928866b3fa0687a6f0763098b27756f1317ad8\": rpc error: code = NotFound desc = could not find container \"5918eddd4d8d7f881902acc813928866b3fa0687a6f0763098b27756f1317ad8\": container with ID starting with 5918eddd4d8d7f881902acc813928866b3fa0687a6f0763098b27756f1317ad8 not found: ID does not exist" Oct 07 14:42:51 crc kubenswrapper[4959]: I1007 14:42:51.146901 4959 scope.go:117] "RemoveContainer" containerID="3eadc7632c0f224744bee6c3e076dcd53e1d7cbfa2f4c07b399d1d4c8a4cdada" Oct 07 14:42:51 crc kubenswrapper[4959]: E1007 14:42:51.147408 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3eadc7632c0f224744bee6c3e076dcd53e1d7cbfa2f4c07b399d1d4c8a4cdada\": container with ID starting with 3eadc7632c0f224744bee6c3e076dcd53e1d7cbfa2f4c07b399d1d4c8a4cdada not found: ID does not exist" containerID="3eadc7632c0f224744bee6c3e076dcd53e1d7cbfa2f4c07b399d1d4c8a4cdada" Oct 07 14:42:51 crc kubenswrapper[4959]: I1007 14:42:51.147437 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3eadc7632c0f224744bee6c3e076dcd53e1d7cbfa2f4c07b399d1d4c8a4cdada"} err="failed to get container status \"3eadc7632c0f224744bee6c3e076dcd53e1d7cbfa2f4c07b399d1d4c8a4cdada\": rpc error: code = NotFound desc = could not find container \"3eadc7632c0f224744bee6c3e076dcd53e1d7cbfa2f4c07b399d1d4c8a4cdada\": container with ID starting with 3eadc7632c0f224744bee6c3e076dcd53e1d7cbfa2f4c07b399d1d4c8a4cdada not found: ID does not exist" Oct 07 14:42:51 crc kubenswrapper[4959]: I1007 14:42:51.147455 4959 scope.go:117] "RemoveContainer" containerID="5918eddd4d8d7f881902acc813928866b3fa0687a6f0763098b27756f1317ad8" Oct 07 14:42:51 crc kubenswrapper[4959]: I1007 14:42:51.147819 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5918eddd4d8d7f881902acc813928866b3fa0687a6f0763098b27756f1317ad8"} err="failed to get container status \"5918eddd4d8d7f881902acc813928866b3fa0687a6f0763098b27756f1317ad8\": rpc error: code = NotFound desc = could not find container \"5918eddd4d8d7f881902acc813928866b3fa0687a6f0763098b27756f1317ad8\": container with ID starting with 5918eddd4d8d7f881902acc813928866b3fa0687a6f0763098b27756f1317ad8 not found: ID does not exist" Oct 07 14:42:51 crc kubenswrapper[4959]: I1007 14:42:51.147852 4959 scope.go:117] "RemoveContainer" containerID="3eadc7632c0f224744bee6c3e076dcd53e1d7cbfa2f4c07b399d1d4c8a4cdada" Oct 07 14:42:51 crc kubenswrapper[4959]: I1007 14:42:51.148234 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3eadc7632c0f224744bee6c3e076dcd53e1d7cbfa2f4c07b399d1d4c8a4cdada"} err="failed to get container status \"3eadc7632c0f224744bee6c3e076dcd53e1d7cbfa2f4c07b399d1d4c8a4cdada\": rpc error: code = NotFound desc = could not find container \"3eadc7632c0f224744bee6c3e076dcd53e1d7cbfa2f4c07b399d1d4c8a4cdada\": container with ID starting with 3eadc7632c0f224744bee6c3e076dcd53e1d7cbfa2f4c07b399d1d4c8a4cdada not found: ID does not exist" Oct 07 14:42:51 crc kubenswrapper[4959]: I1007 14:42:51.151068 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwlqc\" (UniqueName: \"kubernetes.io/projected/34a5b41b-5e32-42e7-a00a-9942fd069d2b-kube-api-access-dwlqc\") pod \"34a5b41b-5e32-42e7-a00a-9942fd069d2b\" (UID: \"34a5b41b-5e32-42e7-a00a-9942fd069d2b\") " Oct 07 14:42:51 crc kubenswrapper[4959]: I1007 14:42:51.162922 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34a5b41b-5e32-42e7-a00a-9942fd069d2b-kube-api-access-dwlqc" (OuterVolumeSpecName: "kube-api-access-dwlqc") pod "34a5b41b-5e32-42e7-a00a-9942fd069d2b" (UID: "34a5b41b-5e32-42e7-a00a-9942fd069d2b"). InnerVolumeSpecName "kube-api-access-dwlqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:42:51 crc kubenswrapper[4959]: I1007 14:42:51.253482 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwlqc\" (UniqueName: \"kubernetes.io/projected/34a5b41b-5e32-42e7-a00a-9942fd069d2b-kube-api-access-dwlqc\") on node \"crc\" DevicePath \"\"" Oct 07 14:42:51 crc kubenswrapper[4959]: I1007 14:42:51.399203 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-57bc4467bb-cdctp"] Oct 07 14:42:51 crc kubenswrapper[4959]: I1007 14:42:51.407718 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-57bc4467bb-cdctp"] Oct 07 14:42:52 crc kubenswrapper[4959]: I1007 14:42:52.668777 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34a5b41b-5e32-42e7-a00a-9942fd069d2b" path="/var/lib/kubelet/pods/34a5b41b-5e32-42e7-a00a-9942fd069d2b/volumes" Oct 07 14:43:14 crc kubenswrapper[4959]: I1007 14:43:14.747772 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pb9mg"] Oct 07 14:43:14 crc kubenswrapper[4959]: E1007 14:43:14.749070 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34a5b41b-5e32-42e7-a00a-9942fd069d2b" containerName="kube-rbac-proxy" Oct 07 14:43:14 crc kubenswrapper[4959]: I1007 14:43:14.749087 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="34a5b41b-5e32-42e7-a00a-9942fd069d2b" containerName="kube-rbac-proxy" Oct 07 14:43:14 crc kubenswrapper[4959]: E1007 14:43:14.749254 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34a5b41b-5e32-42e7-a00a-9942fd069d2b" containerName="operator" Oct 07 14:43:14 crc kubenswrapper[4959]: I1007 14:43:14.749264 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="34a5b41b-5e32-42e7-a00a-9942fd069d2b" containerName="operator" Oct 07 14:43:14 crc kubenswrapper[4959]: I1007 14:43:14.749478 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="34a5b41b-5e32-42e7-a00a-9942fd069d2b" containerName="operator" Oct 07 14:43:14 crc kubenswrapper[4959]: I1007 14:43:14.749507 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="34a5b41b-5e32-42e7-a00a-9942fd069d2b" containerName="kube-rbac-proxy" Oct 07 14:43:14 crc kubenswrapper[4959]: I1007 14:43:14.750958 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pb9mg" Oct 07 14:43:14 crc kubenswrapper[4959]: I1007 14:43:14.768433 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pb9mg"] Oct 07 14:43:14 crc kubenswrapper[4959]: I1007 14:43:14.771988 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6zj4\" (UniqueName: \"kubernetes.io/projected/0d70a844-cd65-44ee-9185-7ed820a128b7-kube-api-access-p6zj4\") pod \"redhat-operators-pb9mg\" (UID: \"0d70a844-cd65-44ee-9185-7ed820a128b7\") " pod="openshift-marketplace/redhat-operators-pb9mg" Oct 07 14:43:14 crc kubenswrapper[4959]: I1007 14:43:14.772110 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d70a844-cd65-44ee-9185-7ed820a128b7-utilities\") pod \"redhat-operators-pb9mg\" (UID: \"0d70a844-cd65-44ee-9185-7ed820a128b7\") " pod="openshift-marketplace/redhat-operators-pb9mg" Oct 07 14:43:14 crc kubenswrapper[4959]: I1007 14:43:14.772156 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d70a844-cd65-44ee-9185-7ed820a128b7-catalog-content\") pod \"redhat-operators-pb9mg\" (UID: \"0d70a844-cd65-44ee-9185-7ed820a128b7\") " pod="openshift-marketplace/redhat-operators-pb9mg" Oct 07 14:43:14 crc kubenswrapper[4959]: I1007 14:43:14.873154 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6zj4\" (UniqueName: \"kubernetes.io/projected/0d70a844-cd65-44ee-9185-7ed820a128b7-kube-api-access-p6zj4\") pod \"redhat-operators-pb9mg\" (UID: \"0d70a844-cd65-44ee-9185-7ed820a128b7\") " pod="openshift-marketplace/redhat-operators-pb9mg" Oct 07 14:43:14 crc kubenswrapper[4959]: I1007 14:43:14.873267 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d70a844-cd65-44ee-9185-7ed820a128b7-utilities\") pod \"redhat-operators-pb9mg\" (UID: \"0d70a844-cd65-44ee-9185-7ed820a128b7\") " pod="openshift-marketplace/redhat-operators-pb9mg" Oct 07 14:43:14 crc kubenswrapper[4959]: I1007 14:43:14.873317 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d70a844-cd65-44ee-9185-7ed820a128b7-catalog-content\") pod \"redhat-operators-pb9mg\" (UID: \"0d70a844-cd65-44ee-9185-7ed820a128b7\") " pod="openshift-marketplace/redhat-operators-pb9mg" Oct 07 14:43:14 crc kubenswrapper[4959]: I1007 14:43:14.873998 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d70a844-cd65-44ee-9185-7ed820a128b7-catalog-content\") pod \"redhat-operators-pb9mg\" (UID: \"0d70a844-cd65-44ee-9185-7ed820a128b7\") " pod="openshift-marketplace/redhat-operators-pb9mg" Oct 07 14:43:14 crc kubenswrapper[4959]: I1007 14:43:14.874007 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d70a844-cd65-44ee-9185-7ed820a128b7-utilities\") pod \"redhat-operators-pb9mg\" (UID: \"0d70a844-cd65-44ee-9185-7ed820a128b7\") " pod="openshift-marketplace/redhat-operators-pb9mg" Oct 07 14:43:14 crc kubenswrapper[4959]: I1007 14:43:14.896050 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6zj4\" (UniqueName: \"kubernetes.io/projected/0d70a844-cd65-44ee-9185-7ed820a128b7-kube-api-access-p6zj4\") pod \"redhat-operators-pb9mg\" (UID: \"0d70a844-cd65-44ee-9185-7ed820a128b7\") " pod="openshift-marketplace/redhat-operators-pb9mg" Oct 07 14:43:15 crc kubenswrapper[4959]: I1007 14:43:15.076213 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pb9mg" Oct 07 14:43:15 crc kubenswrapper[4959]: I1007 14:43:15.570966 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pb9mg"] Oct 07 14:43:16 crc kubenswrapper[4959]: I1007 14:43:16.336215 4959 generic.go:334] "Generic (PLEG): container finished" podID="0d70a844-cd65-44ee-9185-7ed820a128b7" containerID="81405b262bbf5b0fe412601fa0ce33605bb43729482bd14c73f2d2d23088822b" exitCode=0 Oct 07 14:43:16 crc kubenswrapper[4959]: I1007 14:43:16.336274 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pb9mg" event={"ID":"0d70a844-cd65-44ee-9185-7ed820a128b7","Type":"ContainerDied","Data":"81405b262bbf5b0fe412601fa0ce33605bb43729482bd14c73f2d2d23088822b"} Oct 07 14:43:16 crc kubenswrapper[4959]: I1007 14:43:16.336530 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pb9mg" event={"ID":"0d70a844-cd65-44ee-9185-7ed820a128b7","Type":"ContainerStarted","Data":"1479c7576d7f443c27639ad8c01a0a59f6b27912cd8868a1096b026323acb9f4"} Oct 07 14:43:18 crc kubenswrapper[4959]: I1007 14:43:18.357977 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pb9mg" event={"ID":"0d70a844-cd65-44ee-9185-7ed820a128b7","Type":"ContainerStarted","Data":"5183b0a8f731cd31d079986144f928cc046c5d6d5a73396b2925b7f8e6785328"} Oct 07 14:43:21 crc kubenswrapper[4959]: I1007 14:43:21.389397 4959 generic.go:334] "Generic (PLEG): container finished" podID="0d70a844-cd65-44ee-9185-7ed820a128b7" containerID="5183b0a8f731cd31d079986144f928cc046c5d6d5a73396b2925b7f8e6785328" exitCode=0 Oct 07 14:43:21 crc kubenswrapper[4959]: I1007 14:43:21.389504 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pb9mg" event={"ID":"0d70a844-cd65-44ee-9185-7ed820a128b7","Type":"ContainerDied","Data":"5183b0a8f731cd31d079986144f928cc046c5d6d5a73396b2925b7f8e6785328"} Oct 07 14:43:22 crc kubenswrapper[4959]: I1007 14:43:22.399925 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pb9mg" event={"ID":"0d70a844-cd65-44ee-9185-7ed820a128b7","Type":"ContainerStarted","Data":"e3e4a17135063ad2486703178b3ccb9332ab71f3ce7125f30ef8a4c87566f7df"} Oct 07 14:43:22 crc kubenswrapper[4959]: I1007 14:43:22.423926 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pb9mg" podStartSLOduration=2.839888676 podStartE2EDuration="8.423896918s" podCreationTimestamp="2025-10-07 14:43:14 +0000 UTC" firstStartedPulling="2025-10-07 14:43:16.33999525 +0000 UTC m=+3478.423399565" lastFinishedPulling="2025-10-07 14:43:21.924003482 +0000 UTC m=+3484.007407807" observedRunningTime="2025-10-07 14:43:22.420099421 +0000 UTC m=+3484.503503766" watchObservedRunningTime="2025-10-07 14:43:22.423896918 +0000 UTC m=+3484.507301253" Oct 07 14:43:25 crc kubenswrapper[4959]: I1007 14:43:25.076775 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pb9mg" Oct 07 14:43:25 crc kubenswrapper[4959]: I1007 14:43:25.077367 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pb9mg" Oct 07 14:43:26 crc kubenswrapper[4959]: I1007 14:43:26.125724 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pb9mg" podUID="0d70a844-cd65-44ee-9185-7ed820a128b7" containerName="registry-server" probeResult="failure" output=< Oct 07 14:43:26 crc kubenswrapper[4959]: timeout: failed to connect service ":50051" within 1s Oct 07 14:43:26 crc kubenswrapper[4959]: > Oct 07 14:43:27 crc kubenswrapper[4959]: I1007 14:43:27.068534 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-c4f95597f-8ltkv"] Oct 07 14:43:27 crc kubenswrapper[4959]: I1007 14:43:27.070592 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-c4f95597f-8ltkv" Oct 07 14:43:27 crc kubenswrapper[4959]: I1007 14:43:27.082743 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-c4f95597f-8ltkv"] Oct 07 14:43:27 crc kubenswrapper[4959]: I1007 14:43:27.232660 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5rnr\" (UniqueName: \"kubernetes.io/projected/9a28faa3-b29b-468f-af7a-3a6f985e7ce9-kube-api-access-b5rnr\") pod \"test-operator-controller-manager-c4f95597f-8ltkv\" (UID: \"9a28faa3-b29b-468f-af7a-3a6f985e7ce9\") " pod="openstack-operators/test-operator-controller-manager-c4f95597f-8ltkv" Oct 07 14:43:27 crc kubenswrapper[4959]: I1007 14:43:27.335848 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5rnr\" (UniqueName: \"kubernetes.io/projected/9a28faa3-b29b-468f-af7a-3a6f985e7ce9-kube-api-access-b5rnr\") pod \"test-operator-controller-manager-c4f95597f-8ltkv\" (UID: \"9a28faa3-b29b-468f-af7a-3a6f985e7ce9\") " pod="openstack-operators/test-operator-controller-manager-c4f95597f-8ltkv" Oct 07 14:43:27 crc kubenswrapper[4959]: I1007 14:43:27.359819 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5rnr\" (UniqueName: \"kubernetes.io/projected/9a28faa3-b29b-468f-af7a-3a6f985e7ce9-kube-api-access-b5rnr\") pod \"test-operator-controller-manager-c4f95597f-8ltkv\" (UID: \"9a28faa3-b29b-468f-af7a-3a6f985e7ce9\") " pod="openstack-operators/test-operator-controller-manager-c4f95597f-8ltkv" Oct 07 14:43:27 crc kubenswrapper[4959]: I1007 14:43:27.408452 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-c4f95597f-8ltkv" Oct 07 14:43:27 crc kubenswrapper[4959]: I1007 14:43:27.894655 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-c4f95597f-8ltkv"] Oct 07 14:43:28 crc kubenswrapper[4959]: I1007 14:43:28.459710 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-c4f95597f-8ltkv" event={"ID":"9a28faa3-b29b-468f-af7a-3a6f985e7ce9","Type":"ContainerStarted","Data":"26dbf8719d047914dfabfd10a0699951e78e49ce0729ceeeec28d38439a14187"} Oct 07 14:43:29 crc kubenswrapper[4959]: I1007 14:43:29.471943 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-c4f95597f-8ltkv" event={"ID":"9a28faa3-b29b-468f-af7a-3a6f985e7ce9","Type":"ContainerStarted","Data":"e30aa27687ae8664fa56fa75d110090ecce61e41428f9a22f42ac0886a3fb126"} Oct 07 14:43:29 crc kubenswrapper[4959]: I1007 14:43:29.472365 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-c4f95597f-8ltkv" event={"ID":"9a28faa3-b29b-468f-af7a-3a6f985e7ce9","Type":"ContainerStarted","Data":"2d42b016e0ff34fb829c681a3ec694a9105667df2b1826f5f3a90af01f6aca76"} Oct 07 14:43:29 crc kubenswrapper[4959]: I1007 14:43:29.472871 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-c4f95597f-8ltkv" Oct 07 14:43:29 crc kubenswrapper[4959]: I1007 14:43:29.518089 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-c4f95597f-8ltkv" podStartSLOduration=1.395078 podStartE2EDuration="2.518064367s" podCreationTimestamp="2025-10-07 14:43:27 +0000 UTC" firstStartedPulling="2025-10-07 14:43:27.903986621 +0000 UTC m=+3489.987390986" lastFinishedPulling="2025-10-07 14:43:29.026973028 +0000 UTC m=+3491.110377353" observedRunningTime="2025-10-07 14:43:29.512412161 +0000 UTC m=+3491.595816486" watchObservedRunningTime="2025-10-07 14:43:29.518064367 +0000 UTC m=+3491.601468692" Oct 07 14:43:35 crc kubenswrapper[4959]: I1007 14:43:35.127361 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pb9mg" Oct 07 14:43:35 crc kubenswrapper[4959]: I1007 14:43:35.179077 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pb9mg" Oct 07 14:43:35 crc kubenswrapper[4959]: I1007 14:43:35.384010 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pb9mg"] Oct 07 14:43:36 crc kubenswrapper[4959]: I1007 14:43:36.561311 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pb9mg" podUID="0d70a844-cd65-44ee-9185-7ed820a128b7" containerName="registry-server" containerID="cri-o://e3e4a17135063ad2486703178b3ccb9332ab71f3ce7125f30ef8a4c87566f7df" gracePeriod=2 Oct 07 14:43:37 crc kubenswrapper[4959]: I1007 14:43:37.205541 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pb9mg" Oct 07 14:43:37 crc kubenswrapper[4959]: I1007 14:43:37.305246 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d70a844-cd65-44ee-9185-7ed820a128b7-utilities\") pod \"0d70a844-cd65-44ee-9185-7ed820a128b7\" (UID: \"0d70a844-cd65-44ee-9185-7ed820a128b7\") " Oct 07 14:43:37 crc kubenswrapper[4959]: I1007 14:43:37.305350 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6zj4\" (UniqueName: \"kubernetes.io/projected/0d70a844-cd65-44ee-9185-7ed820a128b7-kube-api-access-p6zj4\") pod \"0d70a844-cd65-44ee-9185-7ed820a128b7\" (UID: \"0d70a844-cd65-44ee-9185-7ed820a128b7\") " Oct 07 14:43:37 crc kubenswrapper[4959]: I1007 14:43:37.305474 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d70a844-cd65-44ee-9185-7ed820a128b7-catalog-content\") pod \"0d70a844-cd65-44ee-9185-7ed820a128b7\" (UID: \"0d70a844-cd65-44ee-9185-7ed820a128b7\") " Oct 07 14:43:37 crc kubenswrapper[4959]: I1007 14:43:37.306593 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d70a844-cd65-44ee-9185-7ed820a128b7-utilities" (OuterVolumeSpecName: "utilities") pod "0d70a844-cd65-44ee-9185-7ed820a128b7" (UID: "0d70a844-cd65-44ee-9185-7ed820a128b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:43:37 crc kubenswrapper[4959]: I1007 14:43:37.314309 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d70a844-cd65-44ee-9185-7ed820a128b7-kube-api-access-p6zj4" (OuterVolumeSpecName: "kube-api-access-p6zj4") pod "0d70a844-cd65-44ee-9185-7ed820a128b7" (UID: "0d70a844-cd65-44ee-9185-7ed820a128b7"). InnerVolumeSpecName "kube-api-access-p6zj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:43:37 crc kubenswrapper[4959]: I1007 14:43:37.429394 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d70a844-cd65-44ee-9185-7ed820a128b7-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 14:43:37 crc kubenswrapper[4959]: I1007 14:43:37.429431 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6zj4\" (UniqueName: \"kubernetes.io/projected/0d70a844-cd65-44ee-9185-7ed820a128b7-kube-api-access-p6zj4\") on node \"crc\" DevicePath \"\"" Oct 07 14:43:37 crc kubenswrapper[4959]: I1007 14:43:37.433129 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-c4f95597f-8ltkv" Oct 07 14:43:37 crc kubenswrapper[4959]: I1007 14:43:37.464032 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d70a844-cd65-44ee-9185-7ed820a128b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0d70a844-cd65-44ee-9185-7ed820a128b7" (UID: "0d70a844-cd65-44ee-9185-7ed820a128b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:43:37 crc kubenswrapper[4959]: I1007 14:43:37.496287 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/test-operator-controller-manager-6bb6dcddc-nms79"] Oct 07 14:43:37 crc kubenswrapper[4959]: I1007 14:43:37.496944 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-nms79" podUID="b875d05e-5d50-423f-9c63-66f7f1c2d163" containerName="kube-rbac-proxy" containerID="cri-o://c2b0dd4aef5fb9fe89301676a22a046db6441cc401629e5d243ffaf8fcae95e9" gracePeriod=10 Oct 07 14:43:37 crc kubenswrapper[4959]: I1007 14:43:37.496709 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-nms79" podUID="b875d05e-5d50-423f-9c63-66f7f1c2d163" containerName="manager" containerID="cri-o://cf263d358fab0aa2059ef810b248612afab31b92c1940599c4975776198a917b" gracePeriod=10 Oct 07 14:43:37 crc kubenswrapper[4959]: I1007 14:43:37.532524 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d70a844-cd65-44ee-9185-7ed820a128b7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 14:43:37 crc kubenswrapper[4959]: I1007 14:43:37.574520 4959 generic.go:334] "Generic (PLEG): container finished" podID="0d70a844-cd65-44ee-9185-7ed820a128b7" containerID="e3e4a17135063ad2486703178b3ccb9332ab71f3ce7125f30ef8a4c87566f7df" exitCode=0 Oct 07 14:43:37 crc kubenswrapper[4959]: I1007 14:43:37.574633 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pb9mg" Oct 07 14:43:37 crc kubenswrapper[4959]: I1007 14:43:37.574630 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pb9mg" event={"ID":"0d70a844-cd65-44ee-9185-7ed820a128b7","Type":"ContainerDied","Data":"e3e4a17135063ad2486703178b3ccb9332ab71f3ce7125f30ef8a4c87566f7df"} Oct 07 14:43:37 crc kubenswrapper[4959]: I1007 14:43:37.576017 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pb9mg" event={"ID":"0d70a844-cd65-44ee-9185-7ed820a128b7","Type":"ContainerDied","Data":"1479c7576d7f443c27639ad8c01a0a59f6b27912cd8868a1096b026323acb9f4"} Oct 07 14:43:37 crc kubenswrapper[4959]: I1007 14:43:37.576048 4959 scope.go:117] "RemoveContainer" containerID="e3e4a17135063ad2486703178b3ccb9332ab71f3ce7125f30ef8a4c87566f7df" Oct 07 14:43:37 crc kubenswrapper[4959]: I1007 14:43:37.618411 4959 scope.go:117] "RemoveContainer" containerID="5183b0a8f731cd31d079986144f928cc046c5d6d5a73396b2925b7f8e6785328" Oct 07 14:43:37 crc kubenswrapper[4959]: I1007 14:43:37.630410 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pb9mg"] Oct 07 14:43:37 crc kubenswrapper[4959]: I1007 14:43:37.640362 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pb9mg"] Oct 07 14:43:37 crc kubenswrapper[4959]: I1007 14:43:37.700647 4959 scope.go:117] "RemoveContainer" containerID="81405b262bbf5b0fe412601fa0ce33605bb43729482bd14c73f2d2d23088822b" Oct 07 14:43:37 crc kubenswrapper[4959]: I1007 14:43:37.761824 4959 scope.go:117] "RemoveContainer" containerID="e3e4a17135063ad2486703178b3ccb9332ab71f3ce7125f30ef8a4c87566f7df" Oct 07 14:43:37 crc kubenswrapper[4959]: E1007 14:43:37.762954 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3e4a17135063ad2486703178b3ccb9332ab71f3ce7125f30ef8a4c87566f7df\": container with ID starting with e3e4a17135063ad2486703178b3ccb9332ab71f3ce7125f30ef8a4c87566f7df not found: ID does not exist" containerID="e3e4a17135063ad2486703178b3ccb9332ab71f3ce7125f30ef8a4c87566f7df" Oct 07 14:43:37 crc kubenswrapper[4959]: I1007 14:43:37.763006 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3e4a17135063ad2486703178b3ccb9332ab71f3ce7125f30ef8a4c87566f7df"} err="failed to get container status \"e3e4a17135063ad2486703178b3ccb9332ab71f3ce7125f30ef8a4c87566f7df\": rpc error: code = NotFound desc = could not find container \"e3e4a17135063ad2486703178b3ccb9332ab71f3ce7125f30ef8a4c87566f7df\": container with ID starting with e3e4a17135063ad2486703178b3ccb9332ab71f3ce7125f30ef8a4c87566f7df not found: ID does not exist" Oct 07 14:43:37 crc kubenswrapper[4959]: I1007 14:43:37.763044 4959 scope.go:117] "RemoveContainer" containerID="5183b0a8f731cd31d079986144f928cc046c5d6d5a73396b2925b7f8e6785328" Oct 07 14:43:37 crc kubenswrapper[4959]: E1007 14:43:37.763736 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5183b0a8f731cd31d079986144f928cc046c5d6d5a73396b2925b7f8e6785328\": container with ID starting with 5183b0a8f731cd31d079986144f928cc046c5d6d5a73396b2925b7f8e6785328 not found: ID does not exist" containerID="5183b0a8f731cd31d079986144f928cc046c5d6d5a73396b2925b7f8e6785328" Oct 07 14:43:37 crc kubenswrapper[4959]: I1007 14:43:37.763762 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5183b0a8f731cd31d079986144f928cc046c5d6d5a73396b2925b7f8e6785328"} err="failed to get container status \"5183b0a8f731cd31d079986144f928cc046c5d6d5a73396b2925b7f8e6785328\": rpc error: code = NotFound desc = could not find container \"5183b0a8f731cd31d079986144f928cc046c5d6d5a73396b2925b7f8e6785328\": container with ID starting with 5183b0a8f731cd31d079986144f928cc046c5d6d5a73396b2925b7f8e6785328 not found: ID does not exist" Oct 07 14:43:37 crc kubenswrapper[4959]: I1007 14:43:37.763775 4959 scope.go:117] "RemoveContainer" containerID="81405b262bbf5b0fe412601fa0ce33605bb43729482bd14c73f2d2d23088822b" Oct 07 14:43:37 crc kubenswrapper[4959]: E1007 14:43:37.764085 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81405b262bbf5b0fe412601fa0ce33605bb43729482bd14c73f2d2d23088822b\": container with ID starting with 81405b262bbf5b0fe412601fa0ce33605bb43729482bd14c73f2d2d23088822b not found: ID does not exist" containerID="81405b262bbf5b0fe412601fa0ce33605bb43729482bd14c73f2d2d23088822b" Oct 07 14:43:37 crc kubenswrapper[4959]: I1007 14:43:37.764125 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81405b262bbf5b0fe412601fa0ce33605bb43729482bd14c73f2d2d23088822b"} err="failed to get container status \"81405b262bbf5b0fe412601fa0ce33605bb43729482bd14c73f2d2d23088822b\": rpc error: code = NotFound desc = could not find container \"81405b262bbf5b0fe412601fa0ce33605bb43729482bd14c73f2d2d23088822b\": container with ID starting with 81405b262bbf5b0fe412601fa0ce33605bb43729482bd14c73f2d2d23088822b not found: ID does not exist" Oct 07 14:43:38 crc kubenswrapper[4959]: I1007 14:43:38.004068 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-nms79" Oct 07 14:43:38 crc kubenswrapper[4959]: I1007 14:43:38.052124 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkb64\" (UniqueName: \"kubernetes.io/projected/b875d05e-5d50-423f-9c63-66f7f1c2d163-kube-api-access-kkb64\") pod \"b875d05e-5d50-423f-9c63-66f7f1c2d163\" (UID: \"b875d05e-5d50-423f-9c63-66f7f1c2d163\") " Oct 07 14:43:38 crc kubenswrapper[4959]: I1007 14:43:38.061112 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b875d05e-5d50-423f-9c63-66f7f1c2d163-kube-api-access-kkb64" (OuterVolumeSpecName: "kube-api-access-kkb64") pod "b875d05e-5d50-423f-9c63-66f7f1c2d163" (UID: "b875d05e-5d50-423f-9c63-66f7f1c2d163"). InnerVolumeSpecName "kube-api-access-kkb64". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:43:38 crc kubenswrapper[4959]: I1007 14:43:38.155895 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkb64\" (UniqueName: \"kubernetes.io/projected/b875d05e-5d50-423f-9c63-66f7f1c2d163-kube-api-access-kkb64\") on node \"crc\" DevicePath \"\"" Oct 07 14:43:38 crc kubenswrapper[4959]: I1007 14:43:38.587023 4959 generic.go:334] "Generic (PLEG): container finished" podID="b875d05e-5d50-423f-9c63-66f7f1c2d163" containerID="cf263d358fab0aa2059ef810b248612afab31b92c1940599c4975776198a917b" exitCode=0 Oct 07 14:43:38 crc kubenswrapper[4959]: I1007 14:43:38.588573 4959 generic.go:334] "Generic (PLEG): container finished" podID="b875d05e-5d50-423f-9c63-66f7f1c2d163" containerID="c2b0dd4aef5fb9fe89301676a22a046db6441cc401629e5d243ffaf8fcae95e9" exitCode=0 Oct 07 14:43:38 crc kubenswrapper[4959]: I1007 14:43:38.587388 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-nms79" Oct 07 14:43:38 crc kubenswrapper[4959]: I1007 14:43:38.587354 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-nms79" event={"ID":"b875d05e-5d50-423f-9c63-66f7f1c2d163","Type":"ContainerDied","Data":"cf263d358fab0aa2059ef810b248612afab31b92c1940599c4975776198a917b"} Oct 07 14:43:38 crc kubenswrapper[4959]: I1007 14:43:38.588903 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-nms79" event={"ID":"b875d05e-5d50-423f-9c63-66f7f1c2d163","Type":"ContainerDied","Data":"c2b0dd4aef5fb9fe89301676a22a046db6441cc401629e5d243ffaf8fcae95e9"} Oct 07 14:43:38 crc kubenswrapper[4959]: I1007 14:43:38.588922 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-nms79" event={"ID":"b875d05e-5d50-423f-9c63-66f7f1c2d163","Type":"ContainerDied","Data":"5b04231c1686028ed8e8b2028e67c5f7abc1ce56caf0c1cb7f3410c045e2ec36"} Oct 07 14:43:38 crc kubenswrapper[4959]: I1007 14:43:38.588945 4959 scope.go:117] "RemoveContainer" containerID="cf263d358fab0aa2059ef810b248612afab31b92c1940599c4975776198a917b" Oct 07 14:43:38 crc kubenswrapper[4959]: I1007 14:43:38.623536 4959 scope.go:117] "RemoveContainer" containerID="c2b0dd4aef5fb9fe89301676a22a046db6441cc401629e5d243ffaf8fcae95e9" Oct 07 14:43:38 crc kubenswrapper[4959]: I1007 14:43:38.642753 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/test-operator-controller-manager-6bb6dcddc-nms79"] Oct 07 14:43:38 crc kubenswrapper[4959]: I1007 14:43:38.651949 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/test-operator-controller-manager-6bb6dcddc-nms79"] Oct 07 14:43:38 crc kubenswrapper[4959]: I1007 14:43:38.658719 4959 scope.go:117] "RemoveContainer" containerID="cf263d358fab0aa2059ef810b248612afab31b92c1940599c4975776198a917b" Oct 07 14:43:38 crc kubenswrapper[4959]: E1007 14:43:38.659548 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf263d358fab0aa2059ef810b248612afab31b92c1940599c4975776198a917b\": container with ID starting with cf263d358fab0aa2059ef810b248612afab31b92c1940599c4975776198a917b not found: ID does not exist" containerID="cf263d358fab0aa2059ef810b248612afab31b92c1940599c4975776198a917b" Oct 07 14:43:38 crc kubenswrapper[4959]: I1007 14:43:38.659584 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf263d358fab0aa2059ef810b248612afab31b92c1940599c4975776198a917b"} err="failed to get container status \"cf263d358fab0aa2059ef810b248612afab31b92c1940599c4975776198a917b\": rpc error: code = NotFound desc = could not find container \"cf263d358fab0aa2059ef810b248612afab31b92c1940599c4975776198a917b\": container with ID starting with cf263d358fab0aa2059ef810b248612afab31b92c1940599c4975776198a917b not found: ID does not exist" Oct 07 14:43:38 crc kubenswrapper[4959]: I1007 14:43:38.659615 4959 scope.go:117] "RemoveContainer" containerID="c2b0dd4aef5fb9fe89301676a22a046db6441cc401629e5d243ffaf8fcae95e9" Oct 07 14:43:38 crc kubenswrapper[4959]: E1007 14:43:38.659850 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2b0dd4aef5fb9fe89301676a22a046db6441cc401629e5d243ffaf8fcae95e9\": container with ID starting with c2b0dd4aef5fb9fe89301676a22a046db6441cc401629e5d243ffaf8fcae95e9 not found: ID does not exist" containerID="c2b0dd4aef5fb9fe89301676a22a046db6441cc401629e5d243ffaf8fcae95e9" Oct 07 14:43:38 crc kubenswrapper[4959]: I1007 14:43:38.659874 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2b0dd4aef5fb9fe89301676a22a046db6441cc401629e5d243ffaf8fcae95e9"} err="failed to get container status \"c2b0dd4aef5fb9fe89301676a22a046db6441cc401629e5d243ffaf8fcae95e9\": rpc error: code = NotFound desc = could not find container \"c2b0dd4aef5fb9fe89301676a22a046db6441cc401629e5d243ffaf8fcae95e9\": container with ID starting with c2b0dd4aef5fb9fe89301676a22a046db6441cc401629e5d243ffaf8fcae95e9 not found: ID does not exist" Oct 07 14:43:38 crc kubenswrapper[4959]: I1007 14:43:38.659892 4959 scope.go:117] "RemoveContainer" containerID="cf263d358fab0aa2059ef810b248612afab31b92c1940599c4975776198a917b" Oct 07 14:43:38 crc kubenswrapper[4959]: I1007 14:43:38.660083 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf263d358fab0aa2059ef810b248612afab31b92c1940599c4975776198a917b"} err="failed to get container status \"cf263d358fab0aa2059ef810b248612afab31b92c1940599c4975776198a917b\": rpc error: code = NotFound desc = could not find container \"cf263d358fab0aa2059ef810b248612afab31b92c1940599c4975776198a917b\": container with ID starting with cf263d358fab0aa2059ef810b248612afab31b92c1940599c4975776198a917b not found: ID does not exist" Oct 07 14:43:38 crc kubenswrapper[4959]: I1007 14:43:38.660126 4959 scope.go:117] "RemoveContainer" containerID="c2b0dd4aef5fb9fe89301676a22a046db6441cc401629e5d243ffaf8fcae95e9" Oct 07 14:43:38 crc kubenswrapper[4959]: I1007 14:43:38.660305 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2b0dd4aef5fb9fe89301676a22a046db6441cc401629e5d243ffaf8fcae95e9"} err="failed to get container status \"c2b0dd4aef5fb9fe89301676a22a046db6441cc401629e5d243ffaf8fcae95e9\": rpc error: code = NotFound desc = could not find container \"c2b0dd4aef5fb9fe89301676a22a046db6441cc401629e5d243ffaf8fcae95e9\": container with ID starting with c2b0dd4aef5fb9fe89301676a22a046db6441cc401629e5d243ffaf8fcae95e9 not found: ID does not exist" Oct 07 14:43:38 crc kubenswrapper[4959]: I1007 14:43:38.672394 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d70a844-cd65-44ee-9185-7ed820a128b7" path="/var/lib/kubelet/pods/0d70a844-cd65-44ee-9185-7ed820a128b7/volumes" Oct 07 14:43:38 crc kubenswrapper[4959]: I1007 14:43:38.673198 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b875d05e-5d50-423f-9c63-66f7f1c2d163" path="/var/lib/kubelet/pods/b875d05e-5d50-423f-9c63-66f7f1c2d163/volumes" Oct 07 14:44:01 crc kubenswrapper[4959]: I1007 14:44:01.252954 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-27f25"] Oct 07 14:44:01 crc kubenswrapper[4959]: E1007 14:44:01.253946 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b875d05e-5d50-423f-9c63-66f7f1c2d163" containerName="kube-rbac-proxy" Oct 07 14:44:01 crc kubenswrapper[4959]: I1007 14:44:01.253961 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b875d05e-5d50-423f-9c63-66f7f1c2d163" containerName="kube-rbac-proxy" Oct 07 14:44:01 crc kubenswrapper[4959]: E1007 14:44:01.253970 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d70a844-cd65-44ee-9185-7ed820a128b7" containerName="extract-utilities" Oct 07 14:44:01 crc kubenswrapper[4959]: I1007 14:44:01.253976 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d70a844-cd65-44ee-9185-7ed820a128b7" containerName="extract-utilities" Oct 07 14:44:01 crc kubenswrapper[4959]: E1007 14:44:01.253989 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d70a844-cd65-44ee-9185-7ed820a128b7" containerName="extract-content" Oct 07 14:44:01 crc kubenswrapper[4959]: I1007 14:44:01.253995 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d70a844-cd65-44ee-9185-7ed820a128b7" containerName="extract-content" Oct 07 14:44:01 crc kubenswrapper[4959]: E1007 14:44:01.254015 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d70a844-cd65-44ee-9185-7ed820a128b7" containerName="registry-server" Oct 07 14:44:01 crc kubenswrapper[4959]: I1007 14:44:01.254021 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d70a844-cd65-44ee-9185-7ed820a128b7" containerName="registry-server" Oct 07 14:44:01 crc kubenswrapper[4959]: E1007 14:44:01.254031 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b875d05e-5d50-423f-9c63-66f7f1c2d163" containerName="manager" Oct 07 14:44:01 crc kubenswrapper[4959]: I1007 14:44:01.254037 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b875d05e-5d50-423f-9c63-66f7f1c2d163" containerName="manager" Oct 07 14:44:01 crc kubenswrapper[4959]: I1007 14:44:01.254261 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d70a844-cd65-44ee-9185-7ed820a128b7" containerName="registry-server" Oct 07 14:44:01 crc kubenswrapper[4959]: I1007 14:44:01.254277 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b875d05e-5d50-423f-9c63-66f7f1c2d163" containerName="manager" Oct 07 14:44:01 crc kubenswrapper[4959]: I1007 14:44:01.254299 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b875d05e-5d50-423f-9c63-66f7f1c2d163" containerName="kube-rbac-proxy" Oct 07 14:44:01 crc kubenswrapper[4959]: I1007 14:44:01.255865 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-27f25" Oct 07 14:44:01 crc kubenswrapper[4959]: I1007 14:44:01.267298 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-27f25"] Oct 07 14:44:01 crc kubenswrapper[4959]: I1007 14:44:01.289166 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9-catalog-content\") pod \"redhat-marketplace-27f25\" (UID: \"cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9\") " pod="openshift-marketplace/redhat-marketplace-27f25" Oct 07 14:44:01 crc kubenswrapper[4959]: I1007 14:44:01.289219 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvz9r\" (UniqueName: \"kubernetes.io/projected/cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9-kube-api-access-jvz9r\") pod \"redhat-marketplace-27f25\" (UID: \"cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9\") " pod="openshift-marketplace/redhat-marketplace-27f25" Oct 07 14:44:01 crc kubenswrapper[4959]: I1007 14:44:01.289314 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9-utilities\") pod \"redhat-marketplace-27f25\" (UID: \"cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9\") " pod="openshift-marketplace/redhat-marketplace-27f25" Oct 07 14:44:01 crc kubenswrapper[4959]: I1007 14:44:01.391815 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9-catalog-content\") pod \"redhat-marketplace-27f25\" (UID: \"cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9\") " pod="openshift-marketplace/redhat-marketplace-27f25" Oct 07 14:44:01 crc kubenswrapper[4959]: I1007 14:44:01.391910 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvz9r\" (UniqueName: \"kubernetes.io/projected/cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9-kube-api-access-jvz9r\") pod \"redhat-marketplace-27f25\" (UID: \"cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9\") " pod="openshift-marketplace/redhat-marketplace-27f25" Oct 07 14:44:01 crc kubenswrapper[4959]: I1007 14:44:01.392007 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9-utilities\") pod \"redhat-marketplace-27f25\" (UID: \"cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9\") " pod="openshift-marketplace/redhat-marketplace-27f25" Oct 07 14:44:01 crc kubenswrapper[4959]: I1007 14:44:01.392745 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9-utilities\") pod \"redhat-marketplace-27f25\" (UID: \"cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9\") " pod="openshift-marketplace/redhat-marketplace-27f25" Oct 07 14:44:01 crc kubenswrapper[4959]: I1007 14:44:01.392762 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9-catalog-content\") pod \"redhat-marketplace-27f25\" (UID: \"cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9\") " pod="openshift-marketplace/redhat-marketplace-27f25" Oct 07 14:44:01 crc kubenswrapper[4959]: I1007 14:44:01.414473 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvz9r\" (UniqueName: \"kubernetes.io/projected/cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9-kube-api-access-jvz9r\") pod \"redhat-marketplace-27f25\" (UID: \"cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9\") " pod="openshift-marketplace/redhat-marketplace-27f25" Oct 07 14:44:01 crc kubenswrapper[4959]: I1007 14:44:01.583368 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-27f25" Oct 07 14:44:02 crc kubenswrapper[4959]: I1007 14:44:02.031587 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-27f25"] Oct 07 14:44:02 crc kubenswrapper[4959]: I1007 14:44:02.847586 4959 generic.go:334] "Generic (PLEG): container finished" podID="cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9" containerID="31094014dfad0082b71038c2a180a89caad2bfd44972a18ec5786bca763f663d" exitCode=0 Oct 07 14:44:02 crc kubenswrapper[4959]: I1007 14:44:02.847658 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-27f25" event={"ID":"cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9","Type":"ContainerDied","Data":"31094014dfad0082b71038c2a180a89caad2bfd44972a18ec5786bca763f663d"} Oct 07 14:44:02 crc kubenswrapper[4959]: I1007 14:44:02.849257 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-27f25" event={"ID":"cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9","Type":"ContainerStarted","Data":"af161475e561b73560a050e5f4d37c21ea652f75724a44dad2a85cd2ba6e2903"} Oct 07 14:44:04 crc kubenswrapper[4959]: I1007 14:44:04.883090 4959 generic.go:334] "Generic (PLEG): container finished" podID="cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9" containerID="ba780b67e957d28e988fdba764bad99a618e21f1e4bfca8635d7e0e4216fb334" exitCode=0 Oct 07 14:44:04 crc kubenswrapper[4959]: I1007 14:44:04.883194 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-27f25" event={"ID":"cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9","Type":"ContainerDied","Data":"ba780b67e957d28e988fdba764bad99a618e21f1e4bfca8635d7e0e4216fb334"} Oct 07 14:44:06 crc kubenswrapper[4959]: I1007 14:44:06.904089 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-27f25" event={"ID":"cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9","Type":"ContainerStarted","Data":"673e58d6daf594a838be706587d6eb8ccb70d970c0cb6860bada148da448b6d3"} Oct 07 14:44:06 crc kubenswrapper[4959]: I1007 14:44:06.922500 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-27f25" podStartSLOduration=2.290285098 podStartE2EDuration="5.922477701s" podCreationTimestamp="2025-10-07 14:44:01 +0000 UTC" firstStartedPulling="2025-10-07 14:44:02.850278804 +0000 UTC m=+3524.933683139" lastFinishedPulling="2025-10-07 14:44:06.482471427 +0000 UTC m=+3528.565875742" observedRunningTime="2025-10-07 14:44:06.920620293 +0000 UTC m=+3529.004024618" watchObservedRunningTime="2025-10-07 14:44:06.922477701 +0000 UTC m=+3529.005882026" Oct 07 14:44:11 crc kubenswrapper[4959]: I1007 14:44:11.583942 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-27f25" Oct 07 14:44:11 crc kubenswrapper[4959]: I1007 14:44:11.584555 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-27f25" Oct 07 14:44:11 crc kubenswrapper[4959]: I1007 14:44:11.630362 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-27f25" Oct 07 14:44:12 crc kubenswrapper[4959]: I1007 14:44:12.009334 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-27f25" Oct 07 14:44:12 crc kubenswrapper[4959]: I1007 14:44:12.063589 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-27f25"] Oct 07 14:44:13 crc kubenswrapper[4959]: I1007 14:44:13.985807 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-27f25" podUID="cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9" containerName="registry-server" containerID="cri-o://673e58d6daf594a838be706587d6eb8ccb70d970c0cb6860bada148da448b6d3" gracePeriod=2 Oct 07 14:44:14 crc kubenswrapper[4959]: I1007 14:44:14.467284 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-27f25" Oct 07 14:44:14 crc kubenswrapper[4959]: I1007 14:44:14.635929 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9-utilities\") pod \"cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9\" (UID: \"cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9\") " Oct 07 14:44:14 crc kubenswrapper[4959]: I1007 14:44:14.636416 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9-catalog-content\") pod \"cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9\" (UID: \"cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9\") " Oct 07 14:44:14 crc kubenswrapper[4959]: I1007 14:44:14.636539 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvz9r\" (UniqueName: \"kubernetes.io/projected/cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9-kube-api-access-jvz9r\") pod \"cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9\" (UID: \"cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9\") " Oct 07 14:44:14 crc kubenswrapper[4959]: I1007 14:44:14.637070 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9-utilities" (OuterVolumeSpecName: "utilities") pod "cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9" (UID: "cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:44:14 crc kubenswrapper[4959]: I1007 14:44:14.637323 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 14:44:14 crc kubenswrapper[4959]: I1007 14:44:14.644340 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9-kube-api-access-jvz9r" (OuterVolumeSpecName: "kube-api-access-jvz9r") pod "cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9" (UID: "cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9"). InnerVolumeSpecName "kube-api-access-jvz9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:44:14 crc kubenswrapper[4959]: I1007 14:44:14.649816 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9" (UID: "cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:44:14 crc kubenswrapper[4959]: I1007 14:44:14.740141 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 14:44:14 crc kubenswrapper[4959]: I1007 14:44:14.740198 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvz9r\" (UniqueName: \"kubernetes.io/projected/cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9-kube-api-access-jvz9r\") on node \"crc\" DevicePath \"\"" Oct 07 14:44:14 crc kubenswrapper[4959]: I1007 14:44:14.995596 4959 generic.go:334] "Generic (PLEG): container finished" podID="cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9" containerID="673e58d6daf594a838be706587d6eb8ccb70d970c0cb6860bada148da448b6d3" exitCode=0 Oct 07 14:44:14 crc kubenswrapper[4959]: I1007 14:44:14.995652 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-27f25" event={"ID":"cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9","Type":"ContainerDied","Data":"673e58d6daf594a838be706587d6eb8ccb70d970c0cb6860bada148da448b6d3"} Oct 07 14:44:14 crc kubenswrapper[4959]: I1007 14:44:14.995690 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-27f25" event={"ID":"cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9","Type":"ContainerDied","Data":"af161475e561b73560a050e5f4d37c21ea652f75724a44dad2a85cd2ba6e2903"} Oct 07 14:44:14 crc kubenswrapper[4959]: I1007 14:44:14.995718 4959 scope.go:117] "RemoveContainer" containerID="673e58d6daf594a838be706587d6eb8ccb70d970c0cb6860bada148da448b6d3" Oct 07 14:44:14 crc kubenswrapper[4959]: I1007 14:44:14.995860 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-27f25" Oct 07 14:44:15 crc kubenswrapper[4959]: I1007 14:44:15.018216 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-27f25"] Oct 07 14:44:15 crc kubenswrapper[4959]: I1007 14:44:15.027162 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-27f25"] Oct 07 14:44:15 crc kubenswrapper[4959]: I1007 14:44:15.035306 4959 scope.go:117] "RemoveContainer" containerID="ba780b67e957d28e988fdba764bad99a618e21f1e4bfca8635d7e0e4216fb334" Oct 07 14:44:15 crc kubenswrapper[4959]: I1007 14:44:15.064809 4959 scope.go:117] "RemoveContainer" containerID="31094014dfad0082b71038c2a180a89caad2bfd44972a18ec5786bca763f663d" Oct 07 14:44:15 crc kubenswrapper[4959]: I1007 14:44:15.097034 4959 scope.go:117] "RemoveContainer" containerID="673e58d6daf594a838be706587d6eb8ccb70d970c0cb6860bada148da448b6d3" Oct 07 14:44:15 crc kubenswrapper[4959]: E1007 14:44:15.097881 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"673e58d6daf594a838be706587d6eb8ccb70d970c0cb6860bada148da448b6d3\": container with ID starting with 673e58d6daf594a838be706587d6eb8ccb70d970c0cb6860bada148da448b6d3 not found: ID does not exist" containerID="673e58d6daf594a838be706587d6eb8ccb70d970c0cb6860bada148da448b6d3" Oct 07 14:44:15 crc kubenswrapper[4959]: I1007 14:44:15.097922 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"673e58d6daf594a838be706587d6eb8ccb70d970c0cb6860bada148da448b6d3"} err="failed to get container status \"673e58d6daf594a838be706587d6eb8ccb70d970c0cb6860bada148da448b6d3\": rpc error: code = NotFound desc = could not find container \"673e58d6daf594a838be706587d6eb8ccb70d970c0cb6860bada148da448b6d3\": container with ID starting with 673e58d6daf594a838be706587d6eb8ccb70d970c0cb6860bada148da448b6d3 not found: ID does not exist" Oct 07 14:44:15 crc kubenswrapper[4959]: I1007 14:44:15.097946 4959 scope.go:117] "RemoveContainer" containerID="ba780b67e957d28e988fdba764bad99a618e21f1e4bfca8635d7e0e4216fb334" Oct 07 14:44:15 crc kubenswrapper[4959]: E1007 14:44:15.098281 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba780b67e957d28e988fdba764bad99a618e21f1e4bfca8635d7e0e4216fb334\": container with ID starting with ba780b67e957d28e988fdba764bad99a618e21f1e4bfca8635d7e0e4216fb334 not found: ID does not exist" containerID="ba780b67e957d28e988fdba764bad99a618e21f1e4bfca8635d7e0e4216fb334" Oct 07 14:44:15 crc kubenswrapper[4959]: I1007 14:44:15.098305 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba780b67e957d28e988fdba764bad99a618e21f1e4bfca8635d7e0e4216fb334"} err="failed to get container status \"ba780b67e957d28e988fdba764bad99a618e21f1e4bfca8635d7e0e4216fb334\": rpc error: code = NotFound desc = could not find container \"ba780b67e957d28e988fdba764bad99a618e21f1e4bfca8635d7e0e4216fb334\": container with ID starting with ba780b67e957d28e988fdba764bad99a618e21f1e4bfca8635d7e0e4216fb334 not found: ID does not exist" Oct 07 14:44:15 crc kubenswrapper[4959]: I1007 14:44:15.098316 4959 scope.go:117] "RemoveContainer" containerID="31094014dfad0082b71038c2a180a89caad2bfd44972a18ec5786bca763f663d" Oct 07 14:44:15 crc kubenswrapper[4959]: E1007 14:44:15.098592 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31094014dfad0082b71038c2a180a89caad2bfd44972a18ec5786bca763f663d\": container with ID starting with 31094014dfad0082b71038c2a180a89caad2bfd44972a18ec5786bca763f663d not found: ID does not exist" containerID="31094014dfad0082b71038c2a180a89caad2bfd44972a18ec5786bca763f663d" Oct 07 14:44:15 crc kubenswrapper[4959]: I1007 14:44:15.098646 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31094014dfad0082b71038c2a180a89caad2bfd44972a18ec5786bca763f663d"} err="failed to get container status \"31094014dfad0082b71038c2a180a89caad2bfd44972a18ec5786bca763f663d\": rpc error: code = NotFound desc = could not find container \"31094014dfad0082b71038c2a180a89caad2bfd44972a18ec5786bca763f663d\": container with ID starting with 31094014dfad0082b71038c2a180a89caad2bfd44972a18ec5786bca763f663d not found: ID does not exist" Oct 07 14:44:16 crc kubenswrapper[4959]: I1007 14:44:16.665843 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9" path="/var/lib/kubelet/pods/cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9/volumes" Oct 07 14:44:20 crc kubenswrapper[4959]: I1007 14:44:20.630298 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:44:20 crc kubenswrapper[4959]: I1007 14:44:20.630998 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:44:50 crc kubenswrapper[4959]: I1007 14:44:50.629793 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:44:50 crc kubenswrapper[4959]: I1007 14:44:50.630590 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:45:00 crc kubenswrapper[4959]: I1007 14:45:00.170851 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330805-d5wdj"] Oct 07 14:45:00 crc kubenswrapper[4959]: E1007 14:45:00.172390 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9" containerName="extract-content" Oct 07 14:45:00 crc kubenswrapper[4959]: I1007 14:45:00.172414 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9" containerName="extract-content" Oct 07 14:45:00 crc kubenswrapper[4959]: E1007 14:45:00.172429 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9" containerName="extract-utilities" Oct 07 14:45:00 crc kubenswrapper[4959]: I1007 14:45:00.172439 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9" containerName="extract-utilities" Oct 07 14:45:00 crc kubenswrapper[4959]: E1007 14:45:00.172479 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9" containerName="registry-server" Oct 07 14:45:00 crc kubenswrapper[4959]: I1007 14:45:00.172487 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9" containerName="registry-server" Oct 07 14:45:00 crc kubenswrapper[4959]: I1007 14:45:00.172806 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf3c24e0-9717-4cbd-ab08-1ab8c17e99d9" containerName="registry-server" Oct 07 14:45:00 crc kubenswrapper[4959]: I1007 14:45:00.173816 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330805-d5wdj" Oct 07 14:45:00 crc kubenswrapper[4959]: I1007 14:45:00.176150 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 07 14:45:00 crc kubenswrapper[4959]: I1007 14:45:00.177276 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 07 14:45:00 crc kubenswrapper[4959]: I1007 14:45:00.183992 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330805-d5wdj"] Oct 07 14:45:00 crc kubenswrapper[4959]: I1007 14:45:00.190071 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/097894fe-84e7-4e3b-a7cc-2f8c84026a41-config-volume\") pod \"collect-profiles-29330805-d5wdj\" (UID: \"097894fe-84e7-4e3b-a7cc-2f8c84026a41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330805-d5wdj" Oct 07 14:45:00 crc kubenswrapper[4959]: I1007 14:45:00.190628 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/097894fe-84e7-4e3b-a7cc-2f8c84026a41-secret-volume\") pod \"collect-profiles-29330805-d5wdj\" (UID: \"097894fe-84e7-4e3b-a7cc-2f8c84026a41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330805-d5wdj" Oct 07 14:45:00 crc kubenswrapper[4959]: I1007 14:45:00.190883 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7pzr\" (UniqueName: \"kubernetes.io/projected/097894fe-84e7-4e3b-a7cc-2f8c84026a41-kube-api-access-z7pzr\") pod \"collect-profiles-29330805-d5wdj\" (UID: \"097894fe-84e7-4e3b-a7cc-2f8c84026a41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330805-d5wdj" Oct 07 14:45:00 crc kubenswrapper[4959]: I1007 14:45:00.293927 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7pzr\" (UniqueName: \"kubernetes.io/projected/097894fe-84e7-4e3b-a7cc-2f8c84026a41-kube-api-access-z7pzr\") pod \"collect-profiles-29330805-d5wdj\" (UID: \"097894fe-84e7-4e3b-a7cc-2f8c84026a41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330805-d5wdj" Oct 07 14:45:00 crc kubenswrapper[4959]: I1007 14:45:00.294217 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/097894fe-84e7-4e3b-a7cc-2f8c84026a41-config-volume\") pod \"collect-profiles-29330805-d5wdj\" (UID: \"097894fe-84e7-4e3b-a7cc-2f8c84026a41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330805-d5wdj" Oct 07 14:45:00 crc kubenswrapper[4959]: I1007 14:45:00.294271 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/097894fe-84e7-4e3b-a7cc-2f8c84026a41-secret-volume\") pod \"collect-profiles-29330805-d5wdj\" (UID: \"097894fe-84e7-4e3b-a7cc-2f8c84026a41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330805-d5wdj" Oct 07 14:45:00 crc kubenswrapper[4959]: I1007 14:45:00.295208 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/097894fe-84e7-4e3b-a7cc-2f8c84026a41-config-volume\") pod \"collect-profiles-29330805-d5wdj\" (UID: \"097894fe-84e7-4e3b-a7cc-2f8c84026a41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330805-d5wdj" Oct 07 14:45:00 crc kubenswrapper[4959]: I1007 14:45:00.300321 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/097894fe-84e7-4e3b-a7cc-2f8c84026a41-secret-volume\") pod \"collect-profiles-29330805-d5wdj\" (UID: \"097894fe-84e7-4e3b-a7cc-2f8c84026a41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330805-d5wdj" Oct 07 14:45:00 crc kubenswrapper[4959]: I1007 14:45:00.326821 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7pzr\" (UniqueName: \"kubernetes.io/projected/097894fe-84e7-4e3b-a7cc-2f8c84026a41-kube-api-access-z7pzr\") pod \"collect-profiles-29330805-d5wdj\" (UID: \"097894fe-84e7-4e3b-a7cc-2f8c84026a41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330805-d5wdj" Oct 07 14:45:00 crc kubenswrapper[4959]: I1007 14:45:00.498581 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330805-d5wdj" Oct 07 14:45:01 crc kubenswrapper[4959]: I1007 14:45:01.040810 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330805-d5wdj"] Oct 07 14:45:01 crc kubenswrapper[4959]: I1007 14:45:01.501685 4959 generic.go:334] "Generic (PLEG): container finished" podID="097894fe-84e7-4e3b-a7cc-2f8c84026a41" containerID="21836ce70b60d0a99fa88c251b4a664d70936373bfd9df420d01b00e91d558e1" exitCode=0 Oct 07 14:45:01 crc kubenswrapper[4959]: I1007 14:45:01.501772 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330805-d5wdj" event={"ID":"097894fe-84e7-4e3b-a7cc-2f8c84026a41","Type":"ContainerDied","Data":"21836ce70b60d0a99fa88c251b4a664d70936373bfd9df420d01b00e91d558e1"} Oct 07 14:45:01 crc kubenswrapper[4959]: I1007 14:45:01.502002 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330805-d5wdj" event={"ID":"097894fe-84e7-4e3b-a7cc-2f8c84026a41","Type":"ContainerStarted","Data":"c6deaaa2b149f104e4702ec6a5d6d74a0b9a88443266a2d1cf48acaff478ff7b"} Oct 07 14:45:02 crc kubenswrapper[4959]: I1007 14:45:02.903034 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330805-d5wdj" Oct 07 14:45:02 crc kubenswrapper[4959]: I1007 14:45:02.946078 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/097894fe-84e7-4e3b-a7cc-2f8c84026a41-secret-volume\") pod \"097894fe-84e7-4e3b-a7cc-2f8c84026a41\" (UID: \"097894fe-84e7-4e3b-a7cc-2f8c84026a41\") " Oct 07 14:45:02 crc kubenswrapper[4959]: I1007 14:45:02.946158 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/097894fe-84e7-4e3b-a7cc-2f8c84026a41-config-volume\") pod \"097894fe-84e7-4e3b-a7cc-2f8c84026a41\" (UID: \"097894fe-84e7-4e3b-a7cc-2f8c84026a41\") " Oct 07 14:45:02 crc kubenswrapper[4959]: I1007 14:45:02.946216 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7pzr\" (UniqueName: \"kubernetes.io/projected/097894fe-84e7-4e3b-a7cc-2f8c84026a41-kube-api-access-z7pzr\") pod \"097894fe-84e7-4e3b-a7cc-2f8c84026a41\" (UID: \"097894fe-84e7-4e3b-a7cc-2f8c84026a41\") " Oct 07 14:45:02 crc kubenswrapper[4959]: I1007 14:45:02.946841 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/097894fe-84e7-4e3b-a7cc-2f8c84026a41-config-volume" (OuterVolumeSpecName: "config-volume") pod "097894fe-84e7-4e3b-a7cc-2f8c84026a41" (UID: "097894fe-84e7-4e3b-a7cc-2f8c84026a41"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 14:45:02 crc kubenswrapper[4959]: I1007 14:45:02.959141 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/097894fe-84e7-4e3b-a7cc-2f8c84026a41-kube-api-access-z7pzr" (OuterVolumeSpecName: "kube-api-access-z7pzr") pod "097894fe-84e7-4e3b-a7cc-2f8c84026a41" (UID: "097894fe-84e7-4e3b-a7cc-2f8c84026a41"). InnerVolumeSpecName "kube-api-access-z7pzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:45:02 crc kubenswrapper[4959]: I1007 14:45:02.965445 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/097894fe-84e7-4e3b-a7cc-2f8c84026a41-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "097894fe-84e7-4e3b-a7cc-2f8c84026a41" (UID: "097894fe-84e7-4e3b-a7cc-2f8c84026a41"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 14:45:03 crc kubenswrapper[4959]: I1007 14:45:03.049492 4959 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/097894fe-84e7-4e3b-a7cc-2f8c84026a41-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 07 14:45:03 crc kubenswrapper[4959]: I1007 14:45:03.049539 4959 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/097894fe-84e7-4e3b-a7cc-2f8c84026a41-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 14:45:03 crc kubenswrapper[4959]: I1007 14:45:03.049550 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7pzr\" (UniqueName: \"kubernetes.io/projected/097894fe-84e7-4e3b-a7cc-2f8c84026a41-kube-api-access-z7pzr\") on node \"crc\" DevicePath \"\"" Oct 07 14:45:03 crc kubenswrapper[4959]: I1007 14:45:03.521432 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330805-d5wdj" event={"ID":"097894fe-84e7-4e3b-a7cc-2f8c84026a41","Type":"ContainerDied","Data":"c6deaaa2b149f104e4702ec6a5d6d74a0b9a88443266a2d1cf48acaff478ff7b"} Oct 07 14:45:03 crc kubenswrapper[4959]: I1007 14:45:03.521768 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6deaaa2b149f104e4702ec6a5d6d74a0b9a88443266a2d1cf48acaff478ff7b" Oct 07 14:45:03 crc kubenswrapper[4959]: I1007 14:45:03.521500 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330805-d5wdj" Oct 07 14:45:04 crc kubenswrapper[4959]: I1007 14:45:04.007582 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330760-wn777"] Oct 07 14:45:04 crc kubenswrapper[4959]: I1007 14:45:04.016461 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330760-wn777"] Oct 07 14:45:04 crc kubenswrapper[4959]: I1007 14:45:04.664508 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="444c0979-8132-4868-bd91-5df3278b7660" path="/var/lib/kubelet/pods/444c0979-8132-4868-bd91-5df3278b7660/volumes" Oct 07 14:45:04 crc kubenswrapper[4959]: I1007 14:45:04.886881 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-b9n4q"] Oct 07 14:45:04 crc kubenswrapper[4959]: E1007 14:45:04.887441 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="097894fe-84e7-4e3b-a7cc-2f8c84026a41" containerName="collect-profiles" Oct 07 14:45:04 crc kubenswrapper[4959]: I1007 14:45:04.887460 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="097894fe-84e7-4e3b-a7cc-2f8c84026a41" containerName="collect-profiles" Oct 07 14:45:04 crc kubenswrapper[4959]: I1007 14:45:04.887684 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="097894fe-84e7-4e3b-a7cc-2f8c84026a41" containerName="collect-profiles" Oct 07 14:45:04 crc kubenswrapper[4959]: I1007 14:45:04.889624 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b9n4q" Oct 07 14:45:04 crc kubenswrapper[4959]: I1007 14:45:04.908216 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b9n4q"] Oct 07 14:45:04 crc kubenswrapper[4959]: I1007 14:45:04.986538 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nkjw\" (UniqueName: \"kubernetes.io/projected/3976f21c-2c59-4584-bf59-3d7c40a7e32e-kube-api-access-8nkjw\") pod \"community-operators-b9n4q\" (UID: \"3976f21c-2c59-4584-bf59-3d7c40a7e32e\") " pod="openshift-marketplace/community-operators-b9n4q" Oct 07 14:45:04 crc kubenswrapper[4959]: I1007 14:45:04.986596 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3976f21c-2c59-4584-bf59-3d7c40a7e32e-utilities\") pod \"community-operators-b9n4q\" (UID: \"3976f21c-2c59-4584-bf59-3d7c40a7e32e\") " pod="openshift-marketplace/community-operators-b9n4q" Oct 07 14:45:04 crc kubenswrapper[4959]: I1007 14:45:04.986806 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3976f21c-2c59-4584-bf59-3d7c40a7e32e-catalog-content\") pod \"community-operators-b9n4q\" (UID: \"3976f21c-2c59-4584-bf59-3d7c40a7e32e\") " pod="openshift-marketplace/community-operators-b9n4q" Oct 07 14:45:05 crc kubenswrapper[4959]: I1007 14:45:05.089431 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3976f21c-2c59-4584-bf59-3d7c40a7e32e-catalog-content\") pod \"community-operators-b9n4q\" (UID: \"3976f21c-2c59-4584-bf59-3d7c40a7e32e\") " pod="openshift-marketplace/community-operators-b9n4q" Oct 07 14:45:05 crc kubenswrapper[4959]: I1007 14:45:05.090052 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nkjw\" (UniqueName: \"kubernetes.io/projected/3976f21c-2c59-4584-bf59-3d7c40a7e32e-kube-api-access-8nkjw\") pod \"community-operators-b9n4q\" (UID: \"3976f21c-2c59-4584-bf59-3d7c40a7e32e\") " pod="openshift-marketplace/community-operators-b9n4q" Oct 07 14:45:05 crc kubenswrapper[4959]: I1007 14:45:05.090085 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3976f21c-2c59-4584-bf59-3d7c40a7e32e-utilities\") pod \"community-operators-b9n4q\" (UID: \"3976f21c-2c59-4584-bf59-3d7c40a7e32e\") " pod="openshift-marketplace/community-operators-b9n4q" Oct 07 14:45:05 crc kubenswrapper[4959]: I1007 14:45:05.090241 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3976f21c-2c59-4584-bf59-3d7c40a7e32e-catalog-content\") pod \"community-operators-b9n4q\" (UID: \"3976f21c-2c59-4584-bf59-3d7c40a7e32e\") " pod="openshift-marketplace/community-operators-b9n4q" Oct 07 14:45:05 crc kubenswrapper[4959]: I1007 14:45:05.090729 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3976f21c-2c59-4584-bf59-3d7c40a7e32e-utilities\") pod \"community-operators-b9n4q\" (UID: \"3976f21c-2c59-4584-bf59-3d7c40a7e32e\") " pod="openshift-marketplace/community-operators-b9n4q" Oct 07 14:45:05 crc kubenswrapper[4959]: I1007 14:45:05.114713 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nkjw\" (UniqueName: \"kubernetes.io/projected/3976f21c-2c59-4584-bf59-3d7c40a7e32e-kube-api-access-8nkjw\") pod \"community-operators-b9n4q\" (UID: \"3976f21c-2c59-4584-bf59-3d7c40a7e32e\") " pod="openshift-marketplace/community-operators-b9n4q" Oct 07 14:45:05 crc kubenswrapper[4959]: I1007 14:45:05.217500 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b9n4q" Oct 07 14:45:05 crc kubenswrapper[4959]: I1007 14:45:05.797058 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b9n4q"] Oct 07 14:45:06 crc kubenswrapper[4959]: I1007 14:45:06.552033 4959 generic.go:334] "Generic (PLEG): container finished" podID="3976f21c-2c59-4584-bf59-3d7c40a7e32e" containerID="22afe25e202c67c568555949df8ca7c9107a4b1769c4336675fd21e0c20e787e" exitCode=0 Oct 07 14:45:06 crc kubenswrapper[4959]: I1007 14:45:06.552146 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b9n4q" event={"ID":"3976f21c-2c59-4584-bf59-3d7c40a7e32e","Type":"ContainerDied","Data":"22afe25e202c67c568555949df8ca7c9107a4b1769c4336675fd21e0c20e787e"} Oct 07 14:45:06 crc kubenswrapper[4959]: I1007 14:45:06.552363 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b9n4q" event={"ID":"3976f21c-2c59-4584-bf59-3d7c40a7e32e","Type":"ContainerStarted","Data":"33bce553caf1baba0041d3ae3a2e143562561f4eabae84084ad28efc87f58af5"} Oct 07 14:45:06 crc kubenswrapper[4959]: I1007 14:45:06.554893 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 14:45:08 crc kubenswrapper[4959]: I1007 14:45:08.577793 4959 generic.go:334] "Generic (PLEG): container finished" podID="3976f21c-2c59-4584-bf59-3d7c40a7e32e" containerID="5102982d8376ce01b27b45f28a19ca66bb0c1b9880db7fb070b22f8e99d75c28" exitCode=0 Oct 07 14:45:08 crc kubenswrapper[4959]: I1007 14:45:08.578360 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b9n4q" event={"ID":"3976f21c-2c59-4584-bf59-3d7c40a7e32e","Type":"ContainerDied","Data":"5102982d8376ce01b27b45f28a19ca66bb0c1b9880db7fb070b22f8e99d75c28"} Oct 07 14:45:09 crc kubenswrapper[4959]: I1007 14:45:09.599528 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b9n4q" event={"ID":"3976f21c-2c59-4584-bf59-3d7c40a7e32e","Type":"ContainerStarted","Data":"260d83561fbae96523a2c26f9431fd707f07e24838ef670465a3e17907da37b8"} Oct 07 14:45:09 crc kubenswrapper[4959]: I1007 14:45:09.629333 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-b9n4q" podStartSLOduration=3.202049345 podStartE2EDuration="5.629305432s" podCreationTimestamp="2025-10-07 14:45:04 +0000 UTC" firstStartedPulling="2025-10-07 14:45:06.554618836 +0000 UTC m=+3588.638023161" lastFinishedPulling="2025-10-07 14:45:08.981874923 +0000 UTC m=+3591.065279248" observedRunningTime="2025-10-07 14:45:09.62419596 +0000 UTC m=+3591.707600285" watchObservedRunningTime="2025-10-07 14:45:09.629305432 +0000 UTC m=+3591.712709757" Oct 07 14:45:15 crc kubenswrapper[4959]: I1007 14:45:15.218195 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-b9n4q" Oct 07 14:45:15 crc kubenswrapper[4959]: I1007 14:45:15.218703 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-b9n4q" Oct 07 14:45:15 crc kubenswrapper[4959]: I1007 14:45:15.277314 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-b9n4q" Oct 07 14:45:15 crc kubenswrapper[4959]: I1007 14:45:15.715403 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-b9n4q" Oct 07 14:45:15 crc kubenswrapper[4959]: I1007 14:45:15.762400 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b9n4q"] Oct 07 14:45:17 crc kubenswrapper[4959]: I1007 14:45:17.692741 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-b9n4q" podUID="3976f21c-2c59-4584-bf59-3d7c40a7e32e" containerName="registry-server" containerID="cri-o://260d83561fbae96523a2c26f9431fd707f07e24838ef670465a3e17907da37b8" gracePeriod=2 Oct 07 14:45:18 crc kubenswrapper[4959]: I1007 14:45:18.127274 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b9n4q" Oct 07 14:45:18 crc kubenswrapper[4959]: I1007 14:45:18.304285 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3976f21c-2c59-4584-bf59-3d7c40a7e32e-utilities\") pod \"3976f21c-2c59-4584-bf59-3d7c40a7e32e\" (UID: \"3976f21c-2c59-4584-bf59-3d7c40a7e32e\") " Oct 07 14:45:18 crc kubenswrapper[4959]: I1007 14:45:18.304502 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3976f21c-2c59-4584-bf59-3d7c40a7e32e-catalog-content\") pod \"3976f21c-2c59-4584-bf59-3d7c40a7e32e\" (UID: \"3976f21c-2c59-4584-bf59-3d7c40a7e32e\") " Oct 07 14:45:18 crc kubenswrapper[4959]: I1007 14:45:18.304732 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nkjw\" (UniqueName: \"kubernetes.io/projected/3976f21c-2c59-4584-bf59-3d7c40a7e32e-kube-api-access-8nkjw\") pod \"3976f21c-2c59-4584-bf59-3d7c40a7e32e\" (UID: \"3976f21c-2c59-4584-bf59-3d7c40a7e32e\") " Oct 07 14:45:18 crc kubenswrapper[4959]: I1007 14:45:18.306351 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3976f21c-2c59-4584-bf59-3d7c40a7e32e-utilities" (OuterVolumeSpecName: "utilities") pod "3976f21c-2c59-4584-bf59-3d7c40a7e32e" (UID: "3976f21c-2c59-4584-bf59-3d7c40a7e32e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:45:18 crc kubenswrapper[4959]: I1007 14:45:18.312842 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3976f21c-2c59-4584-bf59-3d7c40a7e32e-kube-api-access-8nkjw" (OuterVolumeSpecName: "kube-api-access-8nkjw") pod "3976f21c-2c59-4584-bf59-3d7c40a7e32e" (UID: "3976f21c-2c59-4584-bf59-3d7c40a7e32e"). InnerVolumeSpecName "kube-api-access-8nkjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:45:18 crc kubenswrapper[4959]: I1007 14:45:18.354687 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3976f21c-2c59-4584-bf59-3d7c40a7e32e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3976f21c-2c59-4584-bf59-3d7c40a7e32e" (UID: "3976f21c-2c59-4584-bf59-3d7c40a7e32e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:45:18 crc kubenswrapper[4959]: I1007 14:45:18.406942 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3976f21c-2c59-4584-bf59-3d7c40a7e32e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 14:45:18 crc kubenswrapper[4959]: I1007 14:45:18.406997 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nkjw\" (UniqueName: \"kubernetes.io/projected/3976f21c-2c59-4584-bf59-3d7c40a7e32e-kube-api-access-8nkjw\") on node \"crc\" DevicePath \"\"" Oct 07 14:45:18 crc kubenswrapper[4959]: I1007 14:45:18.407011 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3976f21c-2c59-4584-bf59-3d7c40a7e32e-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 14:45:18 crc kubenswrapper[4959]: I1007 14:45:18.707646 4959 generic.go:334] "Generic (PLEG): container finished" podID="3976f21c-2c59-4584-bf59-3d7c40a7e32e" containerID="260d83561fbae96523a2c26f9431fd707f07e24838ef670465a3e17907da37b8" exitCode=0 Oct 07 14:45:18 crc kubenswrapper[4959]: I1007 14:45:18.707813 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b9n4q" Oct 07 14:45:18 crc kubenswrapper[4959]: I1007 14:45:18.707935 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b9n4q" event={"ID":"3976f21c-2c59-4584-bf59-3d7c40a7e32e","Type":"ContainerDied","Data":"260d83561fbae96523a2c26f9431fd707f07e24838ef670465a3e17907da37b8"} Oct 07 14:45:18 crc kubenswrapper[4959]: I1007 14:45:18.709546 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b9n4q" event={"ID":"3976f21c-2c59-4584-bf59-3d7c40a7e32e","Type":"ContainerDied","Data":"33bce553caf1baba0041d3ae3a2e143562561f4eabae84084ad28efc87f58af5"} Oct 07 14:45:18 crc kubenswrapper[4959]: I1007 14:45:18.709601 4959 scope.go:117] "RemoveContainer" containerID="260d83561fbae96523a2c26f9431fd707f07e24838ef670465a3e17907da37b8" Oct 07 14:45:18 crc kubenswrapper[4959]: I1007 14:45:18.744729 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b9n4q"] Oct 07 14:45:18 crc kubenswrapper[4959]: I1007 14:45:18.748775 4959 scope.go:117] "RemoveContainer" containerID="5102982d8376ce01b27b45f28a19ca66bb0c1b9880db7fb070b22f8e99d75c28" Oct 07 14:45:18 crc kubenswrapper[4959]: I1007 14:45:18.756569 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-b9n4q"] Oct 07 14:45:18 crc kubenswrapper[4959]: I1007 14:45:18.775842 4959 scope.go:117] "RemoveContainer" containerID="22afe25e202c67c568555949df8ca7c9107a4b1769c4336675fd21e0c20e787e" Oct 07 14:45:18 crc kubenswrapper[4959]: I1007 14:45:18.817544 4959 scope.go:117] "RemoveContainer" containerID="260d83561fbae96523a2c26f9431fd707f07e24838ef670465a3e17907da37b8" Oct 07 14:45:18 crc kubenswrapper[4959]: E1007 14:45:18.818030 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"260d83561fbae96523a2c26f9431fd707f07e24838ef670465a3e17907da37b8\": container with ID starting with 260d83561fbae96523a2c26f9431fd707f07e24838ef670465a3e17907da37b8 not found: ID does not exist" containerID="260d83561fbae96523a2c26f9431fd707f07e24838ef670465a3e17907da37b8" Oct 07 14:45:18 crc kubenswrapper[4959]: I1007 14:45:18.818089 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"260d83561fbae96523a2c26f9431fd707f07e24838ef670465a3e17907da37b8"} err="failed to get container status \"260d83561fbae96523a2c26f9431fd707f07e24838ef670465a3e17907da37b8\": rpc error: code = NotFound desc = could not find container \"260d83561fbae96523a2c26f9431fd707f07e24838ef670465a3e17907da37b8\": container with ID starting with 260d83561fbae96523a2c26f9431fd707f07e24838ef670465a3e17907da37b8 not found: ID does not exist" Oct 07 14:45:18 crc kubenswrapper[4959]: I1007 14:45:18.818153 4959 scope.go:117] "RemoveContainer" containerID="5102982d8376ce01b27b45f28a19ca66bb0c1b9880db7fb070b22f8e99d75c28" Oct 07 14:45:18 crc kubenswrapper[4959]: E1007 14:45:18.818769 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5102982d8376ce01b27b45f28a19ca66bb0c1b9880db7fb070b22f8e99d75c28\": container with ID starting with 5102982d8376ce01b27b45f28a19ca66bb0c1b9880db7fb070b22f8e99d75c28 not found: ID does not exist" containerID="5102982d8376ce01b27b45f28a19ca66bb0c1b9880db7fb070b22f8e99d75c28" Oct 07 14:45:18 crc kubenswrapper[4959]: I1007 14:45:18.818814 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5102982d8376ce01b27b45f28a19ca66bb0c1b9880db7fb070b22f8e99d75c28"} err="failed to get container status \"5102982d8376ce01b27b45f28a19ca66bb0c1b9880db7fb070b22f8e99d75c28\": rpc error: code = NotFound desc = could not find container \"5102982d8376ce01b27b45f28a19ca66bb0c1b9880db7fb070b22f8e99d75c28\": container with ID starting with 5102982d8376ce01b27b45f28a19ca66bb0c1b9880db7fb070b22f8e99d75c28 not found: ID does not exist" Oct 07 14:45:18 crc kubenswrapper[4959]: I1007 14:45:18.818840 4959 scope.go:117] "RemoveContainer" containerID="22afe25e202c67c568555949df8ca7c9107a4b1769c4336675fd21e0c20e787e" Oct 07 14:45:18 crc kubenswrapper[4959]: E1007 14:45:18.819180 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22afe25e202c67c568555949df8ca7c9107a4b1769c4336675fd21e0c20e787e\": container with ID starting with 22afe25e202c67c568555949df8ca7c9107a4b1769c4336675fd21e0c20e787e not found: ID does not exist" containerID="22afe25e202c67c568555949df8ca7c9107a4b1769c4336675fd21e0c20e787e" Oct 07 14:45:18 crc kubenswrapper[4959]: I1007 14:45:18.819219 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22afe25e202c67c568555949df8ca7c9107a4b1769c4336675fd21e0c20e787e"} err="failed to get container status \"22afe25e202c67c568555949df8ca7c9107a4b1769c4336675fd21e0c20e787e\": rpc error: code = NotFound desc = could not find container \"22afe25e202c67c568555949df8ca7c9107a4b1769c4336675fd21e0c20e787e\": container with ID starting with 22afe25e202c67c568555949df8ca7c9107a4b1769c4336675fd21e0c20e787e not found: ID does not exist" Oct 07 14:45:20 crc kubenswrapper[4959]: I1007 14:45:20.630066 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:45:20 crc kubenswrapper[4959]: I1007 14:45:20.630999 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:45:20 crc kubenswrapper[4959]: I1007 14:45:20.631065 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 14:45:20 crc kubenswrapper[4959]: I1007 14:45:20.632292 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f"} pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 14:45:20 crc kubenswrapper[4959]: I1007 14:45:20.632375 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" containerID="cri-o://33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f" gracePeriod=600 Oct 07 14:45:20 crc kubenswrapper[4959]: I1007 14:45:20.667306 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3976f21c-2c59-4584-bf59-3d7c40a7e32e" path="/var/lib/kubelet/pods/3976f21c-2c59-4584-bf59-3d7c40a7e32e/volumes" Oct 07 14:45:20 crc kubenswrapper[4959]: E1007 14:45:20.754802 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:45:21 crc kubenswrapper[4959]: I1007 14:45:21.751049 4959 generic.go:334] "Generic (PLEG): container finished" podID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerID="33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f" exitCode=0 Oct 07 14:45:21 crc kubenswrapper[4959]: I1007 14:45:21.751175 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerDied","Data":"33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f"} Oct 07 14:45:21 crc kubenswrapper[4959]: I1007 14:45:21.751631 4959 scope.go:117] "RemoveContainer" containerID="d95ff1c6b77cf13243ed8e843bbaced75a795144cd83d3ae4f932c4a56aa2beb" Oct 07 14:45:21 crc kubenswrapper[4959]: I1007 14:45:21.752719 4959 scope.go:117] "RemoveContainer" containerID="33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f" Oct 07 14:45:21 crc kubenswrapper[4959]: E1007 14:45:21.753014 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:45:28 crc kubenswrapper[4959]: E1007 14:45:28.549756 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3976f21c_2c59_4584_bf59_3d7c40a7e32e.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3976f21c_2c59_4584_bf59_3d7c40a7e32e.slice/crio-33bce553caf1baba0041d3ae3a2e143562561f4eabae84084ad28efc87f58af5\": RecentStats: unable to find data in memory cache]" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.480983 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest-s00-full"] Oct 07 14:45:29 crc kubenswrapper[4959]: E1007 14:45:29.484802 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3976f21c-2c59-4584-bf59-3d7c40a7e32e" containerName="extract-content" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.485128 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3976f21c-2c59-4584-bf59-3d7c40a7e32e" containerName="extract-content" Oct 07 14:45:29 crc kubenswrapper[4959]: E1007 14:45:29.485231 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3976f21c-2c59-4584-bf59-3d7c40a7e32e" containerName="extract-utilities" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.485306 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3976f21c-2c59-4584-bf59-3d7c40a7e32e" containerName="extract-utilities" Oct 07 14:45:29 crc kubenswrapper[4959]: E1007 14:45:29.485383 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3976f21c-2c59-4584-bf59-3d7c40a7e32e" containerName="registry-server" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.485455 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3976f21c-2c59-4584-bf59-3d7c40a7e32e" containerName="registry-server" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.485846 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="3976f21c-2c59-4584-bf59-3d7c40a7e32e" containerName="registry-server" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.487191 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.490760 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-9g5qz" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.493435 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.494948 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s00-full"] Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.495676 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.495906 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.576662 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-ca-certs\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.576728 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.576814 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.576849 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-ssh-key\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.576881 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-openstack-config\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.576929 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-config-data\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.576960 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-openstack-config-secret\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.576999 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.577020 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-ceph\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.577064 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74sqm\" (UniqueName: \"kubernetes.io/projected/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-kube-api-access-74sqm\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.679589 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-openstack-config-secret\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.679693 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.679725 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-ceph\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.679793 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74sqm\" (UniqueName: \"kubernetes.io/projected/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-kube-api-access-74sqm\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.679845 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-ca-certs\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.679880 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.680005 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.680033 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-ssh-key\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.680064 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-openstack-config\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.680139 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-config-data\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.681546 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-config-data\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.683787 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.683908 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.684233 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.688167 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-openstack-config\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.690780 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-ca-certs\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.692070 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-openstack-config-secret\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.692234 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-ssh-key\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.696599 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-ceph\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.707872 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74sqm\" (UniqueName: \"kubernetes.io/projected/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-kube-api-access-74sqm\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.735067 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest-s00-full\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:29 crc kubenswrapper[4959]: I1007 14:45:29.817978 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-full" Oct 07 14:45:30 crc kubenswrapper[4959]: I1007 14:45:30.371967 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s00-full"] Oct 07 14:45:30 crc kubenswrapper[4959]: I1007 14:45:30.837557 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-full" event={"ID":"b68f21fe-48e8-4b78-8476-ec53a2ca30c8","Type":"ContainerStarted","Data":"c0c64a8bd6af2159fb78c62dce1c6e94ac62e28112e7112d82445a67061eab65"} Oct 07 14:45:31 crc kubenswrapper[4959]: I1007 14:45:31.050485 4959 scope.go:117] "RemoveContainer" containerID="0dab26b0e253a02badce41d64564795611766ab30756fd4692ff9980259c9138" Oct 07 14:45:32 crc kubenswrapper[4959]: I1007 14:45:32.653999 4959 scope.go:117] "RemoveContainer" containerID="33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f" Oct 07 14:45:32 crc kubenswrapper[4959]: E1007 14:45:32.654744 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:45:38 crc kubenswrapper[4959]: E1007 14:45:38.813790 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3976f21c_2c59_4584_bf59_3d7c40a7e32e.slice/crio-33bce553caf1baba0041d3ae3a2e143562561f4eabae84084ad28efc87f58af5\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3976f21c_2c59_4584_bf59_3d7c40a7e32e.slice\": RecentStats: unable to find data in memory cache]" Oct 07 14:45:43 crc kubenswrapper[4959]: I1007 14:45:43.654333 4959 scope.go:117] "RemoveContainer" containerID="33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f" Oct 07 14:45:43 crc kubenswrapper[4959]: E1007 14:45:43.655475 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:45:49 crc kubenswrapper[4959]: E1007 14:45:49.107302 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3976f21c_2c59_4584_bf59_3d7c40a7e32e.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3976f21c_2c59_4584_bf59_3d7c40a7e32e.slice/crio-33bce553caf1baba0041d3ae3a2e143562561f4eabae84084ad28efc87f58af5\": RecentStats: unable to find data in memory cache]" Oct 07 14:45:58 crc kubenswrapper[4959]: I1007 14:45:58.664391 4959 scope.go:117] "RemoveContainer" containerID="33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f" Oct 07 14:45:58 crc kubenswrapper[4959]: E1007 14:45:58.665672 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:45:59 crc kubenswrapper[4959]: E1007 14:45:59.393955 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3976f21c_2c59_4584_bf59_3d7c40a7e32e.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3976f21c_2c59_4584_bf59_3d7c40a7e32e.slice/crio-33bce553caf1baba0041d3ae3a2e143562561f4eabae84084ad28efc87f58af5\": RecentStats: unable to find data in memory cache]" Oct 07 14:46:04 crc kubenswrapper[4959]: E1007 14:46:04.477057 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Oct 07 14:46:04 crc kubenswrapper[4959]: E1007 14:46:04.477804 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceph,ReadOnly:true,MountPath:/etc/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-74sqm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest-s00-full_openstack(b68f21fe-48e8-4b78-8476-ec53a2ca30c8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 14:46:04 crc kubenswrapper[4959]: E1007 14:46:04.479045 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest-s00-full" podUID="b68f21fe-48e8-4b78-8476-ec53a2ca30c8" Oct 07 14:46:05 crc kubenswrapper[4959]: E1007 14:46:05.236728 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest-s00-full" podUID="b68f21fe-48e8-4b78-8476-ec53a2ca30c8" Oct 07 14:46:09 crc kubenswrapper[4959]: E1007 14:46:09.677196 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3976f21c_2c59_4584_bf59_3d7c40a7e32e.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3976f21c_2c59_4584_bf59_3d7c40a7e32e.slice/crio-33bce553caf1baba0041d3ae3a2e143562561f4eabae84084ad28efc87f58af5\": RecentStats: unable to find data in memory cache]" Oct 07 14:46:12 crc kubenswrapper[4959]: I1007 14:46:12.654789 4959 scope.go:117] "RemoveContainer" containerID="33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f" Oct 07 14:46:12 crc kubenswrapper[4959]: E1007 14:46:12.656136 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:46:18 crc kubenswrapper[4959]: I1007 14:46:18.348343 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-full" event={"ID":"b68f21fe-48e8-4b78-8476-ec53a2ca30c8","Type":"ContainerStarted","Data":"7450260130b50a2b82b1c1e913dc57c1afa0f0164a7a1a01cdb7cf5f266fb493"} Oct 07 14:46:18 crc kubenswrapper[4959]: I1007 14:46:18.371365 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest-s00-full" podStartSLOduration=4.330354458 podStartE2EDuration="50.371337437s" podCreationTimestamp="2025-10-07 14:45:28 +0000 UTC" firstStartedPulling="2025-10-07 14:45:30.374464951 +0000 UTC m=+3612.457869266" lastFinishedPulling="2025-10-07 14:46:16.41544792 +0000 UTC m=+3658.498852245" observedRunningTime="2025-10-07 14:46:18.364704537 +0000 UTC m=+3660.448108862" watchObservedRunningTime="2025-10-07 14:46:18.371337437 +0000 UTC m=+3660.454741762" Oct 07 14:46:26 crc kubenswrapper[4959]: I1007 14:46:26.654030 4959 scope.go:117] "RemoveContainer" containerID="33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f" Oct 07 14:46:26 crc kubenswrapper[4959]: E1007 14:46:26.656532 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:46:41 crc kubenswrapper[4959]: I1007 14:46:41.654175 4959 scope.go:117] "RemoveContainer" containerID="33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f" Oct 07 14:46:41 crc kubenswrapper[4959]: E1007 14:46:41.655914 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:46:52 crc kubenswrapper[4959]: I1007 14:46:52.654488 4959 scope.go:117] "RemoveContainer" containerID="33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f" Oct 07 14:46:52 crc kubenswrapper[4959]: E1007 14:46:52.656032 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:47:07 crc kubenswrapper[4959]: I1007 14:47:07.653926 4959 scope.go:117] "RemoveContainer" containerID="33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f" Oct 07 14:47:07 crc kubenswrapper[4959]: E1007 14:47:07.655418 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:47:22 crc kubenswrapper[4959]: I1007 14:47:22.653675 4959 scope.go:117] "RemoveContainer" containerID="33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f" Oct 07 14:47:22 crc kubenswrapper[4959]: E1007 14:47:22.654453 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:47:31 crc kubenswrapper[4959]: I1007 14:47:31.242926 4959 scope.go:117] "RemoveContainer" containerID="31069827fdaa6ac39d0ffc9c949abf0f46ab5f520822a4156363f80d2101047e" Oct 07 14:47:31 crc kubenswrapper[4959]: I1007 14:47:31.273721 4959 scope.go:117] "RemoveContainer" containerID="e160434b947bba7491c28838c32656f376c1a00dd52824b330ed1611ea397eef" Oct 07 14:47:35 crc kubenswrapper[4959]: I1007 14:47:35.654594 4959 scope.go:117] "RemoveContainer" containerID="33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f" Oct 07 14:47:35 crc kubenswrapper[4959]: E1007 14:47:35.655409 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:47:49 crc kubenswrapper[4959]: I1007 14:47:49.824274 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qxt65"] Oct 07 14:47:49 crc kubenswrapper[4959]: I1007 14:47:49.827417 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qxt65" Oct 07 14:47:49 crc kubenswrapper[4959]: I1007 14:47:49.856908 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qxt65"] Oct 07 14:47:49 crc kubenswrapper[4959]: I1007 14:47:49.958291 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef9db97c-567b-48de-82cd-9122da5ac785-utilities\") pod \"certified-operators-qxt65\" (UID: \"ef9db97c-567b-48de-82cd-9122da5ac785\") " pod="openshift-marketplace/certified-operators-qxt65" Oct 07 14:47:49 crc kubenswrapper[4959]: I1007 14:47:49.958347 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwbxb\" (UniqueName: \"kubernetes.io/projected/ef9db97c-567b-48de-82cd-9122da5ac785-kube-api-access-vwbxb\") pod \"certified-operators-qxt65\" (UID: \"ef9db97c-567b-48de-82cd-9122da5ac785\") " pod="openshift-marketplace/certified-operators-qxt65" Oct 07 14:47:49 crc kubenswrapper[4959]: I1007 14:47:49.958640 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef9db97c-567b-48de-82cd-9122da5ac785-catalog-content\") pod \"certified-operators-qxt65\" (UID: \"ef9db97c-567b-48de-82cd-9122da5ac785\") " pod="openshift-marketplace/certified-operators-qxt65" Oct 07 14:47:50 crc kubenswrapper[4959]: I1007 14:47:50.061338 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef9db97c-567b-48de-82cd-9122da5ac785-utilities\") pod \"certified-operators-qxt65\" (UID: \"ef9db97c-567b-48de-82cd-9122da5ac785\") " pod="openshift-marketplace/certified-operators-qxt65" Oct 07 14:47:50 crc kubenswrapper[4959]: I1007 14:47:50.061387 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwbxb\" (UniqueName: \"kubernetes.io/projected/ef9db97c-567b-48de-82cd-9122da5ac785-kube-api-access-vwbxb\") pod \"certified-operators-qxt65\" (UID: \"ef9db97c-567b-48de-82cd-9122da5ac785\") " pod="openshift-marketplace/certified-operators-qxt65" Oct 07 14:47:50 crc kubenswrapper[4959]: I1007 14:47:50.061474 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef9db97c-567b-48de-82cd-9122da5ac785-catalog-content\") pod \"certified-operators-qxt65\" (UID: \"ef9db97c-567b-48de-82cd-9122da5ac785\") " pod="openshift-marketplace/certified-operators-qxt65" Oct 07 14:47:50 crc kubenswrapper[4959]: I1007 14:47:50.061971 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef9db97c-567b-48de-82cd-9122da5ac785-catalog-content\") pod \"certified-operators-qxt65\" (UID: \"ef9db97c-567b-48de-82cd-9122da5ac785\") " pod="openshift-marketplace/certified-operators-qxt65" Oct 07 14:47:50 crc kubenswrapper[4959]: I1007 14:47:50.061980 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef9db97c-567b-48de-82cd-9122da5ac785-utilities\") pod \"certified-operators-qxt65\" (UID: \"ef9db97c-567b-48de-82cd-9122da5ac785\") " pod="openshift-marketplace/certified-operators-qxt65" Oct 07 14:47:50 crc kubenswrapper[4959]: I1007 14:47:50.087464 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwbxb\" (UniqueName: \"kubernetes.io/projected/ef9db97c-567b-48de-82cd-9122da5ac785-kube-api-access-vwbxb\") pod \"certified-operators-qxt65\" (UID: \"ef9db97c-567b-48de-82cd-9122da5ac785\") " pod="openshift-marketplace/certified-operators-qxt65" Oct 07 14:47:50 crc kubenswrapper[4959]: I1007 14:47:50.149287 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qxt65" Oct 07 14:47:50 crc kubenswrapper[4959]: I1007 14:47:50.653944 4959 scope.go:117] "RemoveContainer" containerID="33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f" Oct 07 14:47:50 crc kubenswrapper[4959]: E1007 14:47:50.654640 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:47:50 crc kubenswrapper[4959]: I1007 14:47:50.778417 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qxt65"] Oct 07 14:47:51 crc kubenswrapper[4959]: I1007 14:47:51.395395 4959 generic.go:334] "Generic (PLEG): container finished" podID="ef9db97c-567b-48de-82cd-9122da5ac785" containerID="feef1b508ee96f831d9b20b716da6782d3fdd40fac6b346e0dce8259b7a73dc7" exitCode=0 Oct 07 14:47:51 crc kubenswrapper[4959]: I1007 14:47:51.395508 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qxt65" event={"ID":"ef9db97c-567b-48de-82cd-9122da5ac785","Type":"ContainerDied","Data":"feef1b508ee96f831d9b20b716da6782d3fdd40fac6b346e0dce8259b7a73dc7"} Oct 07 14:47:51 crc kubenswrapper[4959]: I1007 14:47:51.395639 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qxt65" event={"ID":"ef9db97c-567b-48de-82cd-9122da5ac785","Type":"ContainerStarted","Data":"0e86447d688216b44182298fabe7fd9b70b81c9ab4037937af29323fb5fa636c"} Oct 07 14:47:52 crc kubenswrapper[4959]: I1007 14:47:52.421042 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qxt65" event={"ID":"ef9db97c-567b-48de-82cd-9122da5ac785","Type":"ContainerStarted","Data":"6553caca6a6b6471d8767b7997e73e3b946836fc3c7f05dc2f6c1a2c71c3aba8"} Oct 07 14:47:53 crc kubenswrapper[4959]: I1007 14:47:53.433407 4959 generic.go:334] "Generic (PLEG): container finished" podID="ef9db97c-567b-48de-82cd-9122da5ac785" containerID="6553caca6a6b6471d8767b7997e73e3b946836fc3c7f05dc2f6c1a2c71c3aba8" exitCode=0 Oct 07 14:47:53 crc kubenswrapper[4959]: I1007 14:47:53.433511 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qxt65" event={"ID":"ef9db97c-567b-48de-82cd-9122da5ac785","Type":"ContainerDied","Data":"6553caca6a6b6471d8767b7997e73e3b946836fc3c7f05dc2f6c1a2c71c3aba8"} Oct 07 14:47:54 crc kubenswrapper[4959]: I1007 14:47:54.448271 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qxt65" event={"ID":"ef9db97c-567b-48de-82cd-9122da5ac785","Type":"ContainerStarted","Data":"c2937c62b8748b5eb305e8c58e1a0fe7b5f6ece2570e6dadaeef6d285e46da17"} Oct 07 14:47:54 crc kubenswrapper[4959]: I1007 14:47:54.469467 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qxt65" podStartSLOduration=3.008411498 podStartE2EDuration="5.469444174s" podCreationTimestamp="2025-10-07 14:47:49 +0000 UTC" firstStartedPulling="2025-10-07 14:47:51.398491644 +0000 UTC m=+3753.481895969" lastFinishedPulling="2025-10-07 14:47:53.85952432 +0000 UTC m=+3755.942928645" observedRunningTime="2025-10-07 14:47:54.467542885 +0000 UTC m=+3756.550947220" watchObservedRunningTime="2025-10-07 14:47:54.469444174 +0000 UTC m=+3756.552848499" Oct 07 14:48:00 crc kubenswrapper[4959]: I1007 14:48:00.150367 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qxt65" Oct 07 14:48:00 crc kubenswrapper[4959]: I1007 14:48:00.151938 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qxt65" Oct 07 14:48:00 crc kubenswrapper[4959]: I1007 14:48:00.252901 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qxt65" Oct 07 14:48:01 crc kubenswrapper[4959]: I1007 14:48:01.028240 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qxt65" Oct 07 14:48:01 crc kubenswrapper[4959]: I1007 14:48:01.098315 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qxt65"] Oct 07 14:48:02 crc kubenswrapper[4959]: I1007 14:48:02.530637 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qxt65" podUID="ef9db97c-567b-48de-82cd-9122da5ac785" containerName="registry-server" containerID="cri-o://c2937c62b8748b5eb305e8c58e1a0fe7b5f6ece2570e6dadaeef6d285e46da17" gracePeriod=2 Oct 07 14:48:03 crc kubenswrapper[4959]: I1007 14:48:03.123320 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qxt65" Oct 07 14:48:03 crc kubenswrapper[4959]: I1007 14:48:03.236071 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwbxb\" (UniqueName: \"kubernetes.io/projected/ef9db97c-567b-48de-82cd-9122da5ac785-kube-api-access-vwbxb\") pod \"ef9db97c-567b-48de-82cd-9122da5ac785\" (UID: \"ef9db97c-567b-48de-82cd-9122da5ac785\") " Oct 07 14:48:03 crc kubenswrapper[4959]: I1007 14:48:03.236393 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef9db97c-567b-48de-82cd-9122da5ac785-utilities\") pod \"ef9db97c-567b-48de-82cd-9122da5ac785\" (UID: \"ef9db97c-567b-48de-82cd-9122da5ac785\") " Oct 07 14:48:03 crc kubenswrapper[4959]: I1007 14:48:03.236641 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef9db97c-567b-48de-82cd-9122da5ac785-catalog-content\") pod \"ef9db97c-567b-48de-82cd-9122da5ac785\" (UID: \"ef9db97c-567b-48de-82cd-9122da5ac785\") " Oct 07 14:48:03 crc kubenswrapper[4959]: I1007 14:48:03.242682 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef9db97c-567b-48de-82cd-9122da5ac785-kube-api-access-vwbxb" (OuterVolumeSpecName: "kube-api-access-vwbxb") pod "ef9db97c-567b-48de-82cd-9122da5ac785" (UID: "ef9db97c-567b-48de-82cd-9122da5ac785"). InnerVolumeSpecName "kube-api-access-vwbxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:48:03 crc kubenswrapper[4959]: I1007 14:48:03.242800 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef9db97c-567b-48de-82cd-9122da5ac785-utilities" (OuterVolumeSpecName: "utilities") pod "ef9db97c-567b-48de-82cd-9122da5ac785" (UID: "ef9db97c-567b-48de-82cd-9122da5ac785"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:48:03 crc kubenswrapper[4959]: I1007 14:48:03.287389 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef9db97c-567b-48de-82cd-9122da5ac785-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ef9db97c-567b-48de-82cd-9122da5ac785" (UID: "ef9db97c-567b-48de-82cd-9122da5ac785"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:48:03 crc kubenswrapper[4959]: I1007 14:48:03.339937 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef9db97c-567b-48de-82cd-9122da5ac785-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 14:48:03 crc kubenswrapper[4959]: I1007 14:48:03.339976 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwbxb\" (UniqueName: \"kubernetes.io/projected/ef9db97c-567b-48de-82cd-9122da5ac785-kube-api-access-vwbxb\") on node \"crc\" DevicePath \"\"" Oct 07 14:48:03 crc kubenswrapper[4959]: I1007 14:48:03.339986 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef9db97c-567b-48de-82cd-9122da5ac785-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 14:48:03 crc kubenswrapper[4959]: I1007 14:48:03.541344 4959 generic.go:334] "Generic (PLEG): container finished" podID="ef9db97c-567b-48de-82cd-9122da5ac785" containerID="c2937c62b8748b5eb305e8c58e1a0fe7b5f6ece2570e6dadaeef6d285e46da17" exitCode=0 Oct 07 14:48:03 crc kubenswrapper[4959]: I1007 14:48:03.541403 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qxt65" event={"ID":"ef9db97c-567b-48de-82cd-9122da5ac785","Type":"ContainerDied","Data":"c2937c62b8748b5eb305e8c58e1a0fe7b5f6ece2570e6dadaeef6d285e46da17"} Oct 07 14:48:03 crc kubenswrapper[4959]: I1007 14:48:03.541444 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qxt65" event={"ID":"ef9db97c-567b-48de-82cd-9122da5ac785","Type":"ContainerDied","Data":"0e86447d688216b44182298fabe7fd9b70b81c9ab4037937af29323fb5fa636c"} Oct 07 14:48:03 crc kubenswrapper[4959]: I1007 14:48:03.541465 4959 scope.go:117] "RemoveContainer" containerID="c2937c62b8748b5eb305e8c58e1a0fe7b5f6ece2570e6dadaeef6d285e46da17" Oct 07 14:48:03 crc kubenswrapper[4959]: I1007 14:48:03.541478 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qxt65" Oct 07 14:48:03 crc kubenswrapper[4959]: I1007 14:48:03.574946 4959 scope.go:117] "RemoveContainer" containerID="6553caca6a6b6471d8767b7997e73e3b946836fc3c7f05dc2f6c1a2c71c3aba8" Oct 07 14:48:03 crc kubenswrapper[4959]: I1007 14:48:03.604009 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qxt65"] Oct 07 14:48:03 crc kubenswrapper[4959]: I1007 14:48:03.616854 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qxt65"] Oct 07 14:48:03 crc kubenswrapper[4959]: I1007 14:48:03.630797 4959 scope.go:117] "RemoveContainer" containerID="feef1b508ee96f831d9b20b716da6782d3fdd40fac6b346e0dce8259b7a73dc7" Oct 07 14:48:03 crc kubenswrapper[4959]: I1007 14:48:03.668386 4959 scope.go:117] "RemoveContainer" containerID="c2937c62b8748b5eb305e8c58e1a0fe7b5f6ece2570e6dadaeef6d285e46da17" Oct 07 14:48:03 crc kubenswrapper[4959]: E1007 14:48:03.672702 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2937c62b8748b5eb305e8c58e1a0fe7b5f6ece2570e6dadaeef6d285e46da17\": container with ID starting with c2937c62b8748b5eb305e8c58e1a0fe7b5f6ece2570e6dadaeef6d285e46da17 not found: ID does not exist" containerID="c2937c62b8748b5eb305e8c58e1a0fe7b5f6ece2570e6dadaeef6d285e46da17" Oct 07 14:48:03 crc kubenswrapper[4959]: I1007 14:48:03.672747 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2937c62b8748b5eb305e8c58e1a0fe7b5f6ece2570e6dadaeef6d285e46da17"} err="failed to get container status \"c2937c62b8748b5eb305e8c58e1a0fe7b5f6ece2570e6dadaeef6d285e46da17\": rpc error: code = NotFound desc = could not find container \"c2937c62b8748b5eb305e8c58e1a0fe7b5f6ece2570e6dadaeef6d285e46da17\": container with ID starting with c2937c62b8748b5eb305e8c58e1a0fe7b5f6ece2570e6dadaeef6d285e46da17 not found: ID does not exist" Oct 07 14:48:03 crc kubenswrapper[4959]: I1007 14:48:03.672775 4959 scope.go:117] "RemoveContainer" containerID="6553caca6a6b6471d8767b7997e73e3b946836fc3c7f05dc2f6c1a2c71c3aba8" Oct 07 14:48:03 crc kubenswrapper[4959]: E1007 14:48:03.673072 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6553caca6a6b6471d8767b7997e73e3b946836fc3c7f05dc2f6c1a2c71c3aba8\": container with ID starting with 6553caca6a6b6471d8767b7997e73e3b946836fc3c7f05dc2f6c1a2c71c3aba8 not found: ID does not exist" containerID="6553caca6a6b6471d8767b7997e73e3b946836fc3c7f05dc2f6c1a2c71c3aba8" Oct 07 14:48:03 crc kubenswrapper[4959]: I1007 14:48:03.673120 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6553caca6a6b6471d8767b7997e73e3b946836fc3c7f05dc2f6c1a2c71c3aba8"} err="failed to get container status \"6553caca6a6b6471d8767b7997e73e3b946836fc3c7f05dc2f6c1a2c71c3aba8\": rpc error: code = NotFound desc = could not find container \"6553caca6a6b6471d8767b7997e73e3b946836fc3c7f05dc2f6c1a2c71c3aba8\": container with ID starting with 6553caca6a6b6471d8767b7997e73e3b946836fc3c7f05dc2f6c1a2c71c3aba8 not found: ID does not exist" Oct 07 14:48:03 crc kubenswrapper[4959]: I1007 14:48:03.673143 4959 scope.go:117] "RemoveContainer" containerID="feef1b508ee96f831d9b20b716da6782d3fdd40fac6b346e0dce8259b7a73dc7" Oct 07 14:48:03 crc kubenswrapper[4959]: E1007 14:48:03.673390 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"feef1b508ee96f831d9b20b716da6782d3fdd40fac6b346e0dce8259b7a73dc7\": container with ID starting with feef1b508ee96f831d9b20b716da6782d3fdd40fac6b346e0dce8259b7a73dc7 not found: ID does not exist" containerID="feef1b508ee96f831d9b20b716da6782d3fdd40fac6b346e0dce8259b7a73dc7" Oct 07 14:48:03 crc kubenswrapper[4959]: I1007 14:48:03.673413 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"feef1b508ee96f831d9b20b716da6782d3fdd40fac6b346e0dce8259b7a73dc7"} err="failed to get container status \"feef1b508ee96f831d9b20b716da6782d3fdd40fac6b346e0dce8259b7a73dc7\": rpc error: code = NotFound desc = could not find container \"feef1b508ee96f831d9b20b716da6782d3fdd40fac6b346e0dce8259b7a73dc7\": container with ID starting with feef1b508ee96f831d9b20b716da6782d3fdd40fac6b346e0dce8259b7a73dc7 not found: ID does not exist" Oct 07 14:48:04 crc kubenswrapper[4959]: I1007 14:48:04.653899 4959 scope.go:117] "RemoveContainer" containerID="33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f" Oct 07 14:48:04 crc kubenswrapper[4959]: E1007 14:48:04.654606 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:48:04 crc kubenswrapper[4959]: I1007 14:48:04.663399 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef9db97c-567b-48de-82cd-9122da5ac785" path="/var/lib/kubelet/pods/ef9db97c-567b-48de-82cd-9122da5ac785/volumes" Oct 07 14:48:18 crc kubenswrapper[4959]: I1007 14:48:18.655202 4959 scope.go:117] "RemoveContainer" containerID="33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f" Oct 07 14:48:18 crc kubenswrapper[4959]: E1007 14:48:18.655962 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:48:30 crc kubenswrapper[4959]: I1007 14:48:30.655151 4959 scope.go:117] "RemoveContainer" containerID="33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f" Oct 07 14:48:30 crc kubenswrapper[4959]: E1007 14:48:30.656213 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:48:41 crc kubenswrapper[4959]: I1007 14:48:41.653277 4959 scope.go:117] "RemoveContainer" containerID="33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f" Oct 07 14:48:41 crc kubenswrapper[4959]: E1007 14:48:41.655572 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:48:55 crc kubenswrapper[4959]: I1007 14:48:55.654015 4959 scope.go:117] "RemoveContainer" containerID="33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f" Oct 07 14:48:55 crc kubenswrapper[4959]: E1007 14:48:55.656302 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:49:09 crc kubenswrapper[4959]: I1007 14:49:09.663676 4959 scope.go:117] "RemoveContainer" containerID="33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f" Oct 07 14:49:09 crc kubenswrapper[4959]: E1007 14:49:09.668319 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:49:20 crc kubenswrapper[4959]: I1007 14:49:20.653873 4959 scope.go:117] "RemoveContainer" containerID="33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f" Oct 07 14:49:20 crc kubenswrapper[4959]: E1007 14:49:20.657027 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:49:34 crc kubenswrapper[4959]: I1007 14:49:34.653971 4959 scope.go:117] "RemoveContainer" containerID="33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f" Oct 07 14:49:34 crc kubenswrapper[4959]: E1007 14:49:34.657118 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:49:46 crc kubenswrapper[4959]: I1007 14:49:46.654336 4959 scope.go:117] "RemoveContainer" containerID="33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f" Oct 07 14:49:46 crc kubenswrapper[4959]: E1007 14:49:46.655665 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:49:59 crc kubenswrapper[4959]: I1007 14:49:59.653787 4959 scope.go:117] "RemoveContainer" containerID="33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f" Oct 07 14:49:59 crc kubenswrapper[4959]: E1007 14:49:59.654869 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:50:09 crc kubenswrapper[4959]: I1007 14:50:09.070301 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-vnn8v"] Oct 07 14:50:09 crc kubenswrapper[4959]: I1007 14:50:09.080221 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-vnn8v"] Oct 07 14:50:10 crc kubenswrapper[4959]: I1007 14:50:10.676043 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4af5827e-e3e8-41ce-b57f-1ef247587194" path="/var/lib/kubelet/pods/4af5827e-e3e8-41ce-b57f-1ef247587194/volumes" Oct 07 14:50:13 crc kubenswrapper[4959]: I1007 14:50:13.653863 4959 scope.go:117] "RemoveContainer" containerID="33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f" Oct 07 14:50:13 crc kubenswrapper[4959]: E1007 14:50:13.654534 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:50:20 crc kubenswrapper[4959]: I1007 14:50:20.039500 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-09be-account-create-7dklf"] Oct 07 14:50:20 crc kubenswrapper[4959]: I1007 14:50:20.048341 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-09be-account-create-7dklf"] Oct 07 14:50:20 crc kubenswrapper[4959]: I1007 14:50:20.691707 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd865e19-4b6b-4c3f-818a-bd93e7b23045" path="/var/lib/kubelet/pods/cd865e19-4b6b-4c3f-818a-bd93e7b23045/volumes" Oct 07 14:50:25 crc kubenswrapper[4959]: I1007 14:50:25.654020 4959 scope.go:117] "RemoveContainer" containerID="33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f" Oct 07 14:50:25 crc kubenswrapper[4959]: I1007 14:50:25.903951 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerStarted","Data":"4d93ebe2980eadde9513ef305d1bfacd8c628c9df14fa9718e12e94588e90557"} Oct 07 14:50:31 crc kubenswrapper[4959]: I1007 14:50:31.388504 4959 scope.go:117] "RemoveContainer" containerID="dbdec36c647e5432c56d495d0ecde7f51ab8164659ea76c66ca9614967b65478" Oct 07 14:50:31 crc kubenswrapper[4959]: I1007 14:50:31.415249 4959 scope.go:117] "RemoveContainer" containerID="c93a6ebd506351259bbe55103f8ce272d02d62e872c20fb12b4fc2e3cbeaa341" Oct 07 14:50:44 crc kubenswrapper[4959]: I1007 14:50:44.047645 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-bwnpt"] Oct 07 14:50:44 crc kubenswrapper[4959]: I1007 14:50:44.056453 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-bwnpt"] Oct 07 14:50:44 crc kubenswrapper[4959]: I1007 14:50:44.665087 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b129b17a-32c0-4562-bc84-a1e5ea413176" path="/var/lib/kubelet/pods/b129b17a-32c0-4562-bc84-a1e5ea413176/volumes" Oct 07 14:51:31 crc kubenswrapper[4959]: I1007 14:51:31.562798 4959 scope.go:117] "RemoveContainer" containerID="1a85dc16574d9b283a79a3b9aee8744ea227cc83da582fea4869ead30e401f5a" Oct 07 14:52:50 crc kubenswrapper[4959]: I1007 14:52:50.630580 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:52:50 crc kubenswrapper[4959]: I1007 14:52:50.631712 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:53:20 crc kubenswrapper[4959]: I1007 14:53:20.630202 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:53:20 crc kubenswrapper[4959]: I1007 14:53:20.630710 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:53:50 crc kubenswrapper[4959]: I1007 14:53:50.630799 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:53:50 crc kubenswrapper[4959]: I1007 14:53:50.631428 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:53:50 crc kubenswrapper[4959]: I1007 14:53:50.631494 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 14:53:50 crc kubenswrapper[4959]: I1007 14:53:50.632689 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4d93ebe2980eadde9513ef305d1bfacd8c628c9df14fa9718e12e94588e90557"} pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 14:53:50 crc kubenswrapper[4959]: I1007 14:53:50.632774 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" containerID="cri-o://4d93ebe2980eadde9513ef305d1bfacd8c628c9df14fa9718e12e94588e90557" gracePeriod=600 Oct 07 14:53:50 crc kubenswrapper[4959]: I1007 14:53:50.858112 4959 generic.go:334] "Generic (PLEG): container finished" podID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerID="4d93ebe2980eadde9513ef305d1bfacd8c628c9df14fa9718e12e94588e90557" exitCode=0 Oct 07 14:53:50 crc kubenswrapper[4959]: I1007 14:53:50.858183 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerDied","Data":"4d93ebe2980eadde9513ef305d1bfacd8c628c9df14fa9718e12e94588e90557"} Oct 07 14:53:50 crc kubenswrapper[4959]: I1007 14:53:50.858526 4959 scope.go:117] "RemoveContainer" containerID="33a8cecfaa4d1498b06f152b64526e76a31265fa6e19f899cd043b53ea2b742f" Oct 07 14:53:51 crc kubenswrapper[4959]: I1007 14:53:51.869218 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerStarted","Data":"9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204"} Oct 07 14:54:01 crc kubenswrapper[4959]: I1007 14:54:01.323088 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zkptr"] Oct 07 14:54:01 crc kubenswrapper[4959]: E1007 14:54:01.324235 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef9db97c-567b-48de-82cd-9122da5ac785" containerName="extract-utilities" Oct 07 14:54:01 crc kubenswrapper[4959]: I1007 14:54:01.324251 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef9db97c-567b-48de-82cd-9122da5ac785" containerName="extract-utilities" Oct 07 14:54:01 crc kubenswrapper[4959]: E1007 14:54:01.324295 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef9db97c-567b-48de-82cd-9122da5ac785" containerName="extract-content" Oct 07 14:54:01 crc kubenswrapper[4959]: I1007 14:54:01.324305 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef9db97c-567b-48de-82cd-9122da5ac785" containerName="extract-content" Oct 07 14:54:01 crc kubenswrapper[4959]: E1007 14:54:01.324317 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef9db97c-567b-48de-82cd-9122da5ac785" containerName="registry-server" Oct 07 14:54:01 crc kubenswrapper[4959]: I1007 14:54:01.324325 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef9db97c-567b-48de-82cd-9122da5ac785" containerName="registry-server" Oct 07 14:54:01 crc kubenswrapper[4959]: I1007 14:54:01.324663 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef9db97c-567b-48de-82cd-9122da5ac785" containerName="registry-server" Oct 07 14:54:01 crc kubenswrapper[4959]: I1007 14:54:01.326577 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zkptr" Oct 07 14:54:01 crc kubenswrapper[4959]: I1007 14:54:01.349793 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zkptr"] Oct 07 14:54:01 crc kubenswrapper[4959]: I1007 14:54:01.374313 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a64fc48b-c1d1-4ea8-9a2f-7534c54a5287-catalog-content\") pod \"redhat-operators-zkptr\" (UID: \"a64fc48b-c1d1-4ea8-9a2f-7534c54a5287\") " pod="openshift-marketplace/redhat-operators-zkptr" Oct 07 14:54:01 crc kubenswrapper[4959]: I1007 14:54:01.374420 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a64fc48b-c1d1-4ea8-9a2f-7534c54a5287-utilities\") pod \"redhat-operators-zkptr\" (UID: \"a64fc48b-c1d1-4ea8-9a2f-7534c54a5287\") " pod="openshift-marketplace/redhat-operators-zkptr" Oct 07 14:54:01 crc kubenswrapper[4959]: I1007 14:54:01.374709 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nf6x2\" (UniqueName: \"kubernetes.io/projected/a64fc48b-c1d1-4ea8-9a2f-7534c54a5287-kube-api-access-nf6x2\") pod \"redhat-operators-zkptr\" (UID: \"a64fc48b-c1d1-4ea8-9a2f-7534c54a5287\") " pod="openshift-marketplace/redhat-operators-zkptr" Oct 07 14:54:01 crc kubenswrapper[4959]: I1007 14:54:01.476432 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a64fc48b-c1d1-4ea8-9a2f-7534c54a5287-catalog-content\") pod \"redhat-operators-zkptr\" (UID: \"a64fc48b-c1d1-4ea8-9a2f-7534c54a5287\") " pod="openshift-marketplace/redhat-operators-zkptr" Oct 07 14:54:01 crc kubenswrapper[4959]: I1007 14:54:01.476576 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a64fc48b-c1d1-4ea8-9a2f-7534c54a5287-utilities\") pod \"redhat-operators-zkptr\" (UID: \"a64fc48b-c1d1-4ea8-9a2f-7534c54a5287\") " pod="openshift-marketplace/redhat-operators-zkptr" Oct 07 14:54:01 crc kubenswrapper[4959]: I1007 14:54:01.476703 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nf6x2\" (UniqueName: \"kubernetes.io/projected/a64fc48b-c1d1-4ea8-9a2f-7534c54a5287-kube-api-access-nf6x2\") pod \"redhat-operators-zkptr\" (UID: \"a64fc48b-c1d1-4ea8-9a2f-7534c54a5287\") " pod="openshift-marketplace/redhat-operators-zkptr" Oct 07 14:54:01 crc kubenswrapper[4959]: I1007 14:54:01.476950 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a64fc48b-c1d1-4ea8-9a2f-7534c54a5287-catalog-content\") pod \"redhat-operators-zkptr\" (UID: \"a64fc48b-c1d1-4ea8-9a2f-7534c54a5287\") " pod="openshift-marketplace/redhat-operators-zkptr" Oct 07 14:54:01 crc kubenswrapper[4959]: I1007 14:54:01.477215 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a64fc48b-c1d1-4ea8-9a2f-7534c54a5287-utilities\") pod \"redhat-operators-zkptr\" (UID: \"a64fc48b-c1d1-4ea8-9a2f-7534c54a5287\") " pod="openshift-marketplace/redhat-operators-zkptr" Oct 07 14:54:01 crc kubenswrapper[4959]: I1007 14:54:01.510131 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nf6x2\" (UniqueName: \"kubernetes.io/projected/a64fc48b-c1d1-4ea8-9a2f-7534c54a5287-kube-api-access-nf6x2\") pod \"redhat-operators-zkptr\" (UID: \"a64fc48b-c1d1-4ea8-9a2f-7534c54a5287\") " pod="openshift-marketplace/redhat-operators-zkptr" Oct 07 14:54:01 crc kubenswrapper[4959]: I1007 14:54:01.660601 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zkptr" Oct 07 14:54:02 crc kubenswrapper[4959]: I1007 14:54:02.179058 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zkptr"] Oct 07 14:54:03 crc kubenswrapper[4959]: I1007 14:54:03.018498 4959 generic.go:334] "Generic (PLEG): container finished" podID="a64fc48b-c1d1-4ea8-9a2f-7534c54a5287" containerID="81f0dc2ec232925c404e5d6d16130af80f4169aa9325dee1294eb81f484252dd" exitCode=0 Oct 07 14:54:03 crc kubenswrapper[4959]: I1007 14:54:03.018590 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zkptr" event={"ID":"a64fc48b-c1d1-4ea8-9a2f-7534c54a5287","Type":"ContainerDied","Data":"81f0dc2ec232925c404e5d6d16130af80f4169aa9325dee1294eb81f484252dd"} Oct 07 14:54:03 crc kubenswrapper[4959]: I1007 14:54:03.019138 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zkptr" event={"ID":"a64fc48b-c1d1-4ea8-9a2f-7534c54a5287","Type":"ContainerStarted","Data":"3c56f1bcb356e9d5c17f206791ffd8f12403b6c0a5c3897ec177717393f02e93"} Oct 07 14:54:03 crc kubenswrapper[4959]: I1007 14:54:03.021780 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 14:54:05 crc kubenswrapper[4959]: I1007 14:54:05.039022 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zkptr" event={"ID":"a64fc48b-c1d1-4ea8-9a2f-7534c54a5287","Type":"ContainerStarted","Data":"68776345e17e3672570a361773d44f3f484ea6c172aed2bcc5ace3e511bb707f"} Oct 07 14:54:20 crc kubenswrapper[4959]: I1007 14:54:20.203489 4959 generic.go:334] "Generic (PLEG): container finished" podID="a64fc48b-c1d1-4ea8-9a2f-7534c54a5287" containerID="68776345e17e3672570a361773d44f3f484ea6c172aed2bcc5ace3e511bb707f" exitCode=0 Oct 07 14:54:20 crc kubenswrapper[4959]: I1007 14:54:20.203582 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zkptr" event={"ID":"a64fc48b-c1d1-4ea8-9a2f-7534c54a5287","Type":"ContainerDied","Data":"68776345e17e3672570a361773d44f3f484ea6c172aed2bcc5ace3e511bb707f"} Oct 07 14:54:22 crc kubenswrapper[4959]: I1007 14:54:22.236216 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zkptr" event={"ID":"a64fc48b-c1d1-4ea8-9a2f-7534c54a5287","Type":"ContainerStarted","Data":"3f9a3e77768813889f2ac3707d5385e82e2a4d3d0c0749c182ce04fb57d257e8"} Oct 07 14:54:22 crc kubenswrapper[4959]: I1007 14:54:22.270183 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zkptr" podStartSLOduration=3.203060846 podStartE2EDuration="21.270154219s" podCreationTimestamp="2025-10-07 14:54:01 +0000 UTC" firstStartedPulling="2025-10-07 14:54:03.021534349 +0000 UTC m=+4125.104938674" lastFinishedPulling="2025-10-07 14:54:21.088627722 +0000 UTC m=+4143.172032047" observedRunningTime="2025-10-07 14:54:22.264778542 +0000 UTC m=+4144.348182877" watchObservedRunningTime="2025-10-07 14:54:22.270154219 +0000 UTC m=+4144.353558544" Oct 07 14:54:29 crc kubenswrapper[4959]: I1007 14:54:29.547243 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wp89g"] Oct 07 14:54:29 crc kubenswrapper[4959]: I1007 14:54:29.550892 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wp89g" Oct 07 14:54:29 crc kubenswrapper[4959]: I1007 14:54:29.556753 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wp89g"] Oct 07 14:54:29 crc kubenswrapper[4959]: I1007 14:54:29.654907 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/558fe3f7-f896-4032-9c18-c9cb43beba3d-catalog-content\") pod \"redhat-marketplace-wp89g\" (UID: \"558fe3f7-f896-4032-9c18-c9cb43beba3d\") " pod="openshift-marketplace/redhat-marketplace-wp89g" Oct 07 14:54:29 crc kubenswrapper[4959]: I1007 14:54:29.655048 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/558fe3f7-f896-4032-9c18-c9cb43beba3d-utilities\") pod \"redhat-marketplace-wp89g\" (UID: \"558fe3f7-f896-4032-9c18-c9cb43beba3d\") " pod="openshift-marketplace/redhat-marketplace-wp89g" Oct 07 14:54:29 crc kubenswrapper[4959]: I1007 14:54:29.655391 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5r8f\" (UniqueName: \"kubernetes.io/projected/558fe3f7-f896-4032-9c18-c9cb43beba3d-kube-api-access-x5r8f\") pod \"redhat-marketplace-wp89g\" (UID: \"558fe3f7-f896-4032-9c18-c9cb43beba3d\") " pod="openshift-marketplace/redhat-marketplace-wp89g" Oct 07 14:54:29 crc kubenswrapper[4959]: I1007 14:54:29.758027 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/558fe3f7-f896-4032-9c18-c9cb43beba3d-catalog-content\") pod \"redhat-marketplace-wp89g\" (UID: \"558fe3f7-f896-4032-9c18-c9cb43beba3d\") " pod="openshift-marketplace/redhat-marketplace-wp89g" Oct 07 14:54:29 crc kubenswrapper[4959]: I1007 14:54:29.758197 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/558fe3f7-f896-4032-9c18-c9cb43beba3d-utilities\") pod \"redhat-marketplace-wp89g\" (UID: \"558fe3f7-f896-4032-9c18-c9cb43beba3d\") " pod="openshift-marketplace/redhat-marketplace-wp89g" Oct 07 14:54:29 crc kubenswrapper[4959]: I1007 14:54:29.758752 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/558fe3f7-f896-4032-9c18-c9cb43beba3d-catalog-content\") pod \"redhat-marketplace-wp89g\" (UID: \"558fe3f7-f896-4032-9c18-c9cb43beba3d\") " pod="openshift-marketplace/redhat-marketplace-wp89g" Oct 07 14:54:29 crc kubenswrapper[4959]: I1007 14:54:29.758783 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/558fe3f7-f896-4032-9c18-c9cb43beba3d-utilities\") pod \"redhat-marketplace-wp89g\" (UID: \"558fe3f7-f896-4032-9c18-c9cb43beba3d\") " pod="openshift-marketplace/redhat-marketplace-wp89g" Oct 07 14:54:29 crc kubenswrapper[4959]: I1007 14:54:29.759047 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5r8f\" (UniqueName: \"kubernetes.io/projected/558fe3f7-f896-4032-9c18-c9cb43beba3d-kube-api-access-x5r8f\") pod \"redhat-marketplace-wp89g\" (UID: \"558fe3f7-f896-4032-9c18-c9cb43beba3d\") " pod="openshift-marketplace/redhat-marketplace-wp89g" Oct 07 14:54:29 crc kubenswrapper[4959]: I1007 14:54:29.785412 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5r8f\" (UniqueName: \"kubernetes.io/projected/558fe3f7-f896-4032-9c18-c9cb43beba3d-kube-api-access-x5r8f\") pod \"redhat-marketplace-wp89g\" (UID: \"558fe3f7-f896-4032-9c18-c9cb43beba3d\") " pod="openshift-marketplace/redhat-marketplace-wp89g" Oct 07 14:54:29 crc kubenswrapper[4959]: I1007 14:54:29.877412 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wp89g" Oct 07 14:54:30 crc kubenswrapper[4959]: I1007 14:54:30.437559 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wp89g"] Oct 07 14:54:30 crc kubenswrapper[4959]: W1007 14:54:30.449563 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod558fe3f7_f896_4032_9c18_c9cb43beba3d.slice/crio-003ed81717b242ee7555bb83b68e26602ec33decb62045685f50ed5cbf857be6 WatchSource:0}: Error finding container 003ed81717b242ee7555bb83b68e26602ec33decb62045685f50ed5cbf857be6: Status 404 returned error can't find the container with id 003ed81717b242ee7555bb83b68e26602ec33decb62045685f50ed5cbf857be6 Oct 07 14:54:31 crc kubenswrapper[4959]: I1007 14:54:31.337623 4959 generic.go:334] "Generic (PLEG): container finished" podID="558fe3f7-f896-4032-9c18-c9cb43beba3d" containerID="45643148a048fd9c614cd9bc4a3a550d07e4962f6a00d64e79d5ebb9f189471f" exitCode=0 Oct 07 14:54:31 crc kubenswrapper[4959]: I1007 14:54:31.337718 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wp89g" event={"ID":"558fe3f7-f896-4032-9c18-c9cb43beba3d","Type":"ContainerDied","Data":"45643148a048fd9c614cd9bc4a3a550d07e4962f6a00d64e79d5ebb9f189471f"} Oct 07 14:54:31 crc kubenswrapper[4959]: I1007 14:54:31.338289 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wp89g" event={"ID":"558fe3f7-f896-4032-9c18-c9cb43beba3d","Type":"ContainerStarted","Data":"003ed81717b242ee7555bb83b68e26602ec33decb62045685f50ed5cbf857be6"} Oct 07 14:54:31 crc kubenswrapper[4959]: I1007 14:54:31.661381 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zkptr" Oct 07 14:54:31 crc kubenswrapper[4959]: I1007 14:54:31.661435 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zkptr" Oct 07 14:54:31 crc kubenswrapper[4959]: I1007 14:54:31.713629 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zkptr" Oct 07 14:54:32 crc kubenswrapper[4959]: I1007 14:54:32.406117 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zkptr" Oct 07 14:54:33 crc kubenswrapper[4959]: I1007 14:54:33.510481 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zkptr"] Oct 07 14:54:34 crc kubenswrapper[4959]: I1007 14:54:34.370143 4959 generic.go:334] "Generic (PLEG): container finished" podID="558fe3f7-f896-4032-9c18-c9cb43beba3d" containerID="f3f0052b2d339bb3f19a53a2c08d0eaeff8a35c381dcb20bd5bfd44be9d098b9" exitCode=0 Oct 07 14:54:34 crc kubenswrapper[4959]: I1007 14:54:34.370208 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wp89g" event={"ID":"558fe3f7-f896-4032-9c18-c9cb43beba3d","Type":"ContainerDied","Data":"f3f0052b2d339bb3f19a53a2c08d0eaeff8a35c381dcb20bd5bfd44be9d098b9"} Oct 07 14:54:34 crc kubenswrapper[4959]: I1007 14:54:34.370411 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zkptr" podUID="a64fc48b-c1d1-4ea8-9a2f-7534c54a5287" containerName="registry-server" containerID="cri-o://3f9a3e77768813889f2ac3707d5385e82e2a4d3d0c0749c182ce04fb57d257e8" gracePeriod=2 Oct 07 14:54:35 crc kubenswrapper[4959]: I1007 14:54:35.012587 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zkptr" Oct 07 14:54:35 crc kubenswrapper[4959]: I1007 14:54:35.084998 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a64fc48b-c1d1-4ea8-9a2f-7534c54a5287-catalog-content\") pod \"a64fc48b-c1d1-4ea8-9a2f-7534c54a5287\" (UID: \"a64fc48b-c1d1-4ea8-9a2f-7534c54a5287\") " Oct 07 14:54:35 crc kubenswrapper[4959]: I1007 14:54:35.085262 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nf6x2\" (UniqueName: \"kubernetes.io/projected/a64fc48b-c1d1-4ea8-9a2f-7534c54a5287-kube-api-access-nf6x2\") pod \"a64fc48b-c1d1-4ea8-9a2f-7534c54a5287\" (UID: \"a64fc48b-c1d1-4ea8-9a2f-7534c54a5287\") " Oct 07 14:54:35 crc kubenswrapper[4959]: I1007 14:54:35.085340 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a64fc48b-c1d1-4ea8-9a2f-7534c54a5287-utilities\") pod \"a64fc48b-c1d1-4ea8-9a2f-7534c54a5287\" (UID: \"a64fc48b-c1d1-4ea8-9a2f-7534c54a5287\") " Oct 07 14:54:35 crc kubenswrapper[4959]: I1007 14:54:35.086251 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a64fc48b-c1d1-4ea8-9a2f-7534c54a5287-utilities" (OuterVolumeSpecName: "utilities") pod "a64fc48b-c1d1-4ea8-9a2f-7534c54a5287" (UID: "a64fc48b-c1d1-4ea8-9a2f-7534c54a5287"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:54:35 crc kubenswrapper[4959]: I1007 14:54:35.093462 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a64fc48b-c1d1-4ea8-9a2f-7534c54a5287-kube-api-access-nf6x2" (OuterVolumeSpecName: "kube-api-access-nf6x2") pod "a64fc48b-c1d1-4ea8-9a2f-7534c54a5287" (UID: "a64fc48b-c1d1-4ea8-9a2f-7534c54a5287"). InnerVolumeSpecName "kube-api-access-nf6x2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:54:35 crc kubenswrapper[4959]: I1007 14:54:35.184738 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a64fc48b-c1d1-4ea8-9a2f-7534c54a5287-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a64fc48b-c1d1-4ea8-9a2f-7534c54a5287" (UID: "a64fc48b-c1d1-4ea8-9a2f-7534c54a5287"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:54:35 crc kubenswrapper[4959]: I1007 14:54:35.188414 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a64fc48b-c1d1-4ea8-9a2f-7534c54a5287-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 14:54:35 crc kubenswrapper[4959]: I1007 14:54:35.188445 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nf6x2\" (UniqueName: \"kubernetes.io/projected/a64fc48b-c1d1-4ea8-9a2f-7534c54a5287-kube-api-access-nf6x2\") on node \"crc\" DevicePath \"\"" Oct 07 14:54:35 crc kubenswrapper[4959]: I1007 14:54:35.188460 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a64fc48b-c1d1-4ea8-9a2f-7534c54a5287-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 14:54:35 crc kubenswrapper[4959]: I1007 14:54:35.389216 4959 generic.go:334] "Generic (PLEG): container finished" podID="a64fc48b-c1d1-4ea8-9a2f-7534c54a5287" containerID="3f9a3e77768813889f2ac3707d5385e82e2a4d3d0c0749c182ce04fb57d257e8" exitCode=0 Oct 07 14:54:35 crc kubenswrapper[4959]: I1007 14:54:35.389281 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zkptr" event={"ID":"a64fc48b-c1d1-4ea8-9a2f-7534c54a5287","Type":"ContainerDied","Data":"3f9a3e77768813889f2ac3707d5385e82e2a4d3d0c0749c182ce04fb57d257e8"} Oct 07 14:54:35 crc kubenswrapper[4959]: I1007 14:54:35.389349 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zkptr" Oct 07 14:54:35 crc kubenswrapper[4959]: I1007 14:54:35.389362 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zkptr" event={"ID":"a64fc48b-c1d1-4ea8-9a2f-7534c54a5287","Type":"ContainerDied","Data":"3c56f1bcb356e9d5c17f206791ffd8f12403b6c0a5c3897ec177717393f02e93"} Oct 07 14:54:35 crc kubenswrapper[4959]: I1007 14:54:35.389404 4959 scope.go:117] "RemoveContainer" containerID="3f9a3e77768813889f2ac3707d5385e82e2a4d3d0c0749c182ce04fb57d257e8" Oct 07 14:54:35 crc kubenswrapper[4959]: I1007 14:54:35.428207 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zkptr"] Oct 07 14:54:35 crc kubenswrapper[4959]: I1007 14:54:35.436202 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zkptr"] Oct 07 14:54:35 crc kubenswrapper[4959]: I1007 14:54:35.668127 4959 scope.go:117] "RemoveContainer" containerID="68776345e17e3672570a361773d44f3f484ea6c172aed2bcc5ace3e511bb707f" Oct 07 14:54:35 crc kubenswrapper[4959]: I1007 14:54:35.699190 4959 scope.go:117] "RemoveContainer" containerID="81f0dc2ec232925c404e5d6d16130af80f4169aa9325dee1294eb81f484252dd" Oct 07 14:54:35 crc kubenswrapper[4959]: I1007 14:54:35.904120 4959 scope.go:117] "RemoveContainer" containerID="3f9a3e77768813889f2ac3707d5385e82e2a4d3d0c0749c182ce04fb57d257e8" Oct 07 14:54:35 crc kubenswrapper[4959]: E1007 14:54:35.904980 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f9a3e77768813889f2ac3707d5385e82e2a4d3d0c0749c182ce04fb57d257e8\": container with ID starting with 3f9a3e77768813889f2ac3707d5385e82e2a4d3d0c0749c182ce04fb57d257e8 not found: ID does not exist" containerID="3f9a3e77768813889f2ac3707d5385e82e2a4d3d0c0749c182ce04fb57d257e8" Oct 07 14:54:35 crc kubenswrapper[4959]: I1007 14:54:35.905022 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f9a3e77768813889f2ac3707d5385e82e2a4d3d0c0749c182ce04fb57d257e8"} err="failed to get container status \"3f9a3e77768813889f2ac3707d5385e82e2a4d3d0c0749c182ce04fb57d257e8\": rpc error: code = NotFound desc = could not find container \"3f9a3e77768813889f2ac3707d5385e82e2a4d3d0c0749c182ce04fb57d257e8\": container with ID starting with 3f9a3e77768813889f2ac3707d5385e82e2a4d3d0c0749c182ce04fb57d257e8 not found: ID does not exist" Oct 07 14:54:35 crc kubenswrapper[4959]: I1007 14:54:35.905050 4959 scope.go:117] "RemoveContainer" containerID="68776345e17e3672570a361773d44f3f484ea6c172aed2bcc5ace3e511bb707f" Oct 07 14:54:35 crc kubenswrapper[4959]: E1007 14:54:35.905360 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68776345e17e3672570a361773d44f3f484ea6c172aed2bcc5ace3e511bb707f\": container with ID starting with 68776345e17e3672570a361773d44f3f484ea6c172aed2bcc5ace3e511bb707f not found: ID does not exist" containerID="68776345e17e3672570a361773d44f3f484ea6c172aed2bcc5ace3e511bb707f" Oct 07 14:54:35 crc kubenswrapper[4959]: I1007 14:54:35.905395 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68776345e17e3672570a361773d44f3f484ea6c172aed2bcc5ace3e511bb707f"} err="failed to get container status \"68776345e17e3672570a361773d44f3f484ea6c172aed2bcc5ace3e511bb707f\": rpc error: code = NotFound desc = could not find container \"68776345e17e3672570a361773d44f3f484ea6c172aed2bcc5ace3e511bb707f\": container with ID starting with 68776345e17e3672570a361773d44f3f484ea6c172aed2bcc5ace3e511bb707f not found: ID does not exist" Oct 07 14:54:35 crc kubenswrapper[4959]: I1007 14:54:35.905607 4959 scope.go:117] "RemoveContainer" containerID="81f0dc2ec232925c404e5d6d16130af80f4169aa9325dee1294eb81f484252dd" Oct 07 14:54:35 crc kubenswrapper[4959]: E1007 14:54:35.905979 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81f0dc2ec232925c404e5d6d16130af80f4169aa9325dee1294eb81f484252dd\": container with ID starting with 81f0dc2ec232925c404e5d6d16130af80f4169aa9325dee1294eb81f484252dd not found: ID does not exist" containerID="81f0dc2ec232925c404e5d6d16130af80f4169aa9325dee1294eb81f484252dd" Oct 07 14:54:35 crc kubenswrapper[4959]: I1007 14:54:35.906060 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81f0dc2ec232925c404e5d6d16130af80f4169aa9325dee1294eb81f484252dd"} err="failed to get container status \"81f0dc2ec232925c404e5d6d16130af80f4169aa9325dee1294eb81f484252dd\": rpc error: code = NotFound desc = could not find container \"81f0dc2ec232925c404e5d6d16130af80f4169aa9325dee1294eb81f484252dd\": container with ID starting with 81f0dc2ec232925c404e5d6d16130af80f4169aa9325dee1294eb81f484252dd not found: ID does not exist" Oct 07 14:54:36 crc kubenswrapper[4959]: I1007 14:54:36.664244 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a64fc48b-c1d1-4ea8-9a2f-7534c54a5287" path="/var/lib/kubelet/pods/a64fc48b-c1d1-4ea8-9a2f-7534c54a5287/volumes" Oct 07 14:54:38 crc kubenswrapper[4959]: I1007 14:54:38.443691 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wp89g" event={"ID":"558fe3f7-f896-4032-9c18-c9cb43beba3d","Type":"ContainerStarted","Data":"e2384dbc5c262faf7f4952e21f18532646e63dab7494634f342f928928645e0d"} Oct 07 14:54:38 crc kubenswrapper[4959]: I1007 14:54:38.462843 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wp89g" podStartSLOduration=3.742994296 podStartE2EDuration="9.462827257s" podCreationTimestamp="2025-10-07 14:54:29 +0000 UTC" firstStartedPulling="2025-10-07 14:54:31.343083635 +0000 UTC m=+4153.426487960" lastFinishedPulling="2025-10-07 14:54:37.062916586 +0000 UTC m=+4159.146320921" observedRunningTime="2025-10-07 14:54:38.460513658 +0000 UTC m=+4160.543917973" watchObservedRunningTime="2025-10-07 14:54:38.462827257 +0000 UTC m=+4160.546231582" Oct 07 14:54:39 crc kubenswrapper[4959]: I1007 14:54:39.878508 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wp89g" Oct 07 14:54:39 crc kubenswrapper[4959]: I1007 14:54:39.879140 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wp89g" Oct 07 14:54:39 crc kubenswrapper[4959]: I1007 14:54:39.928807 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wp89g" Oct 07 14:54:50 crc kubenswrapper[4959]: I1007 14:54:50.022864 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wp89g" Oct 07 14:54:50 crc kubenswrapper[4959]: I1007 14:54:50.078820 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wp89g"] Oct 07 14:54:50 crc kubenswrapper[4959]: I1007 14:54:50.546823 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wp89g" podUID="558fe3f7-f896-4032-9c18-c9cb43beba3d" containerName="registry-server" containerID="cri-o://e2384dbc5c262faf7f4952e21f18532646e63dab7494634f342f928928645e0d" gracePeriod=2 Oct 07 14:54:51 crc kubenswrapper[4959]: I1007 14:54:51.193670 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wp89g" Oct 07 14:54:51 crc kubenswrapper[4959]: I1007 14:54:51.338290 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/558fe3f7-f896-4032-9c18-c9cb43beba3d-catalog-content\") pod \"558fe3f7-f896-4032-9c18-c9cb43beba3d\" (UID: \"558fe3f7-f896-4032-9c18-c9cb43beba3d\") " Oct 07 14:54:51 crc kubenswrapper[4959]: I1007 14:54:51.338482 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5r8f\" (UniqueName: \"kubernetes.io/projected/558fe3f7-f896-4032-9c18-c9cb43beba3d-kube-api-access-x5r8f\") pod \"558fe3f7-f896-4032-9c18-c9cb43beba3d\" (UID: \"558fe3f7-f896-4032-9c18-c9cb43beba3d\") " Oct 07 14:54:51 crc kubenswrapper[4959]: I1007 14:54:51.338527 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/558fe3f7-f896-4032-9c18-c9cb43beba3d-utilities\") pod \"558fe3f7-f896-4032-9c18-c9cb43beba3d\" (UID: \"558fe3f7-f896-4032-9c18-c9cb43beba3d\") " Oct 07 14:54:51 crc kubenswrapper[4959]: I1007 14:54:51.339733 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/558fe3f7-f896-4032-9c18-c9cb43beba3d-utilities" (OuterVolumeSpecName: "utilities") pod "558fe3f7-f896-4032-9c18-c9cb43beba3d" (UID: "558fe3f7-f896-4032-9c18-c9cb43beba3d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:54:51 crc kubenswrapper[4959]: I1007 14:54:51.353597 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/558fe3f7-f896-4032-9c18-c9cb43beba3d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "558fe3f7-f896-4032-9c18-c9cb43beba3d" (UID: "558fe3f7-f896-4032-9c18-c9cb43beba3d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:54:51 crc kubenswrapper[4959]: I1007 14:54:51.440879 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/558fe3f7-f896-4032-9c18-c9cb43beba3d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 14:54:51 crc kubenswrapper[4959]: I1007 14:54:51.440920 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/558fe3f7-f896-4032-9c18-c9cb43beba3d-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 14:54:51 crc kubenswrapper[4959]: I1007 14:54:51.581157 4959 generic.go:334] "Generic (PLEG): container finished" podID="558fe3f7-f896-4032-9c18-c9cb43beba3d" containerID="e2384dbc5c262faf7f4952e21f18532646e63dab7494634f342f928928645e0d" exitCode=0 Oct 07 14:54:51 crc kubenswrapper[4959]: I1007 14:54:51.581212 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wp89g" event={"ID":"558fe3f7-f896-4032-9c18-c9cb43beba3d","Type":"ContainerDied","Data":"e2384dbc5c262faf7f4952e21f18532646e63dab7494634f342f928928645e0d"} Oct 07 14:54:51 crc kubenswrapper[4959]: I1007 14:54:51.581224 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wp89g" Oct 07 14:54:51 crc kubenswrapper[4959]: I1007 14:54:51.581245 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wp89g" event={"ID":"558fe3f7-f896-4032-9c18-c9cb43beba3d","Type":"ContainerDied","Data":"003ed81717b242ee7555bb83b68e26602ec33decb62045685f50ed5cbf857be6"} Oct 07 14:54:51 crc kubenswrapper[4959]: I1007 14:54:51.581269 4959 scope.go:117] "RemoveContainer" containerID="e2384dbc5c262faf7f4952e21f18532646e63dab7494634f342f928928645e0d" Oct 07 14:54:51 crc kubenswrapper[4959]: I1007 14:54:51.600746 4959 scope.go:117] "RemoveContainer" containerID="f3f0052b2d339bb3f19a53a2c08d0eaeff8a35c381dcb20bd5bfd44be9d098b9" Oct 07 14:54:51 crc kubenswrapper[4959]: I1007 14:54:51.875351 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/558fe3f7-f896-4032-9c18-c9cb43beba3d-kube-api-access-x5r8f" (OuterVolumeSpecName: "kube-api-access-x5r8f") pod "558fe3f7-f896-4032-9c18-c9cb43beba3d" (UID: "558fe3f7-f896-4032-9c18-c9cb43beba3d"). InnerVolumeSpecName "kube-api-access-x5r8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:54:51 crc kubenswrapper[4959]: I1007 14:54:51.893347 4959 scope.go:117] "RemoveContainer" containerID="45643148a048fd9c614cd9bc4a3a550d07e4962f6a00d64e79d5ebb9f189471f" Oct 07 14:54:51 crc kubenswrapper[4959]: I1007 14:54:51.951925 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5r8f\" (UniqueName: \"kubernetes.io/projected/558fe3f7-f896-4032-9c18-c9cb43beba3d-kube-api-access-x5r8f\") on node \"crc\" DevicePath \"\"" Oct 07 14:54:51 crc kubenswrapper[4959]: I1007 14:54:51.979790 4959 scope.go:117] "RemoveContainer" containerID="e2384dbc5c262faf7f4952e21f18532646e63dab7494634f342f928928645e0d" Oct 07 14:54:51 crc kubenswrapper[4959]: E1007 14:54:51.980391 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2384dbc5c262faf7f4952e21f18532646e63dab7494634f342f928928645e0d\": container with ID starting with e2384dbc5c262faf7f4952e21f18532646e63dab7494634f342f928928645e0d not found: ID does not exist" containerID="e2384dbc5c262faf7f4952e21f18532646e63dab7494634f342f928928645e0d" Oct 07 14:54:51 crc kubenswrapper[4959]: I1007 14:54:51.980424 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2384dbc5c262faf7f4952e21f18532646e63dab7494634f342f928928645e0d"} err="failed to get container status \"e2384dbc5c262faf7f4952e21f18532646e63dab7494634f342f928928645e0d\": rpc error: code = NotFound desc = could not find container \"e2384dbc5c262faf7f4952e21f18532646e63dab7494634f342f928928645e0d\": container with ID starting with e2384dbc5c262faf7f4952e21f18532646e63dab7494634f342f928928645e0d not found: ID does not exist" Oct 07 14:54:51 crc kubenswrapper[4959]: I1007 14:54:51.980447 4959 scope.go:117] "RemoveContainer" containerID="f3f0052b2d339bb3f19a53a2c08d0eaeff8a35c381dcb20bd5bfd44be9d098b9" Oct 07 14:54:51 crc kubenswrapper[4959]: E1007 14:54:51.981443 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3f0052b2d339bb3f19a53a2c08d0eaeff8a35c381dcb20bd5bfd44be9d098b9\": container with ID starting with f3f0052b2d339bb3f19a53a2c08d0eaeff8a35c381dcb20bd5bfd44be9d098b9 not found: ID does not exist" containerID="f3f0052b2d339bb3f19a53a2c08d0eaeff8a35c381dcb20bd5bfd44be9d098b9" Oct 07 14:54:51 crc kubenswrapper[4959]: I1007 14:54:51.981464 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3f0052b2d339bb3f19a53a2c08d0eaeff8a35c381dcb20bd5bfd44be9d098b9"} err="failed to get container status \"f3f0052b2d339bb3f19a53a2c08d0eaeff8a35c381dcb20bd5bfd44be9d098b9\": rpc error: code = NotFound desc = could not find container \"f3f0052b2d339bb3f19a53a2c08d0eaeff8a35c381dcb20bd5bfd44be9d098b9\": container with ID starting with f3f0052b2d339bb3f19a53a2c08d0eaeff8a35c381dcb20bd5bfd44be9d098b9 not found: ID does not exist" Oct 07 14:54:51 crc kubenswrapper[4959]: I1007 14:54:51.981479 4959 scope.go:117] "RemoveContainer" containerID="45643148a048fd9c614cd9bc4a3a550d07e4962f6a00d64e79d5ebb9f189471f" Oct 07 14:54:51 crc kubenswrapper[4959]: E1007 14:54:51.981885 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45643148a048fd9c614cd9bc4a3a550d07e4962f6a00d64e79d5ebb9f189471f\": container with ID starting with 45643148a048fd9c614cd9bc4a3a550d07e4962f6a00d64e79d5ebb9f189471f not found: ID does not exist" containerID="45643148a048fd9c614cd9bc4a3a550d07e4962f6a00d64e79d5ebb9f189471f" Oct 07 14:54:51 crc kubenswrapper[4959]: I1007 14:54:51.981908 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45643148a048fd9c614cd9bc4a3a550d07e4962f6a00d64e79d5ebb9f189471f"} err="failed to get container status \"45643148a048fd9c614cd9bc4a3a550d07e4962f6a00d64e79d5ebb9f189471f\": rpc error: code = NotFound desc = could not find container \"45643148a048fd9c614cd9bc4a3a550d07e4962f6a00d64e79d5ebb9f189471f\": container with ID starting with 45643148a048fd9c614cd9bc4a3a550d07e4962f6a00d64e79d5ebb9f189471f not found: ID does not exist" Oct 07 14:54:51 crc kubenswrapper[4959]: I1007 14:54:51.986893 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wp89g"] Oct 07 14:54:51 crc kubenswrapper[4959]: I1007 14:54:51.997602 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wp89g"] Oct 07 14:54:52 crc kubenswrapper[4959]: I1007 14:54:52.664900 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="558fe3f7-f896-4032-9c18-c9cb43beba3d" path="/var/lib/kubelet/pods/558fe3f7-f896-4032-9c18-c9cb43beba3d/volumes" Oct 07 14:56:18 crc kubenswrapper[4959]: I1007 14:56:18.424259 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fbmpt"] Oct 07 14:56:18 crc kubenswrapper[4959]: E1007 14:56:18.425496 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a64fc48b-c1d1-4ea8-9a2f-7534c54a5287" containerName="registry-server" Oct 07 14:56:18 crc kubenswrapper[4959]: I1007 14:56:18.425512 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a64fc48b-c1d1-4ea8-9a2f-7534c54a5287" containerName="registry-server" Oct 07 14:56:18 crc kubenswrapper[4959]: E1007 14:56:18.425528 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a64fc48b-c1d1-4ea8-9a2f-7534c54a5287" containerName="extract-content" Oct 07 14:56:18 crc kubenswrapper[4959]: I1007 14:56:18.425534 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a64fc48b-c1d1-4ea8-9a2f-7534c54a5287" containerName="extract-content" Oct 07 14:56:18 crc kubenswrapper[4959]: E1007 14:56:18.425551 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="558fe3f7-f896-4032-9c18-c9cb43beba3d" containerName="registry-server" Oct 07 14:56:18 crc kubenswrapper[4959]: I1007 14:56:18.425558 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="558fe3f7-f896-4032-9c18-c9cb43beba3d" containerName="registry-server" Oct 07 14:56:18 crc kubenswrapper[4959]: E1007 14:56:18.425583 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="558fe3f7-f896-4032-9c18-c9cb43beba3d" containerName="extract-content" Oct 07 14:56:18 crc kubenswrapper[4959]: I1007 14:56:18.425589 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="558fe3f7-f896-4032-9c18-c9cb43beba3d" containerName="extract-content" Oct 07 14:56:18 crc kubenswrapper[4959]: E1007 14:56:18.425602 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="558fe3f7-f896-4032-9c18-c9cb43beba3d" containerName="extract-utilities" Oct 07 14:56:18 crc kubenswrapper[4959]: I1007 14:56:18.425612 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="558fe3f7-f896-4032-9c18-c9cb43beba3d" containerName="extract-utilities" Oct 07 14:56:18 crc kubenswrapper[4959]: E1007 14:56:18.425623 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a64fc48b-c1d1-4ea8-9a2f-7534c54a5287" containerName="extract-utilities" Oct 07 14:56:18 crc kubenswrapper[4959]: I1007 14:56:18.425630 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a64fc48b-c1d1-4ea8-9a2f-7534c54a5287" containerName="extract-utilities" Oct 07 14:56:18 crc kubenswrapper[4959]: I1007 14:56:18.425892 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a64fc48b-c1d1-4ea8-9a2f-7534c54a5287" containerName="registry-server" Oct 07 14:56:18 crc kubenswrapper[4959]: I1007 14:56:18.425906 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="558fe3f7-f896-4032-9c18-c9cb43beba3d" containerName="registry-server" Oct 07 14:56:18 crc kubenswrapper[4959]: I1007 14:56:18.427691 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fbmpt" Oct 07 14:56:18 crc kubenswrapper[4959]: I1007 14:56:18.436891 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fbmpt"] Oct 07 14:56:18 crc kubenswrapper[4959]: I1007 14:56:18.491971 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71912490-8062-40d6-b950-e81b6142d40c-utilities\") pod \"community-operators-fbmpt\" (UID: \"71912490-8062-40d6-b950-e81b6142d40c\") " pod="openshift-marketplace/community-operators-fbmpt" Oct 07 14:56:18 crc kubenswrapper[4959]: I1007 14:56:18.492233 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71912490-8062-40d6-b950-e81b6142d40c-catalog-content\") pod \"community-operators-fbmpt\" (UID: \"71912490-8062-40d6-b950-e81b6142d40c\") " pod="openshift-marketplace/community-operators-fbmpt" Oct 07 14:56:18 crc kubenswrapper[4959]: I1007 14:56:18.492556 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmncz\" (UniqueName: \"kubernetes.io/projected/71912490-8062-40d6-b950-e81b6142d40c-kube-api-access-hmncz\") pod \"community-operators-fbmpt\" (UID: \"71912490-8062-40d6-b950-e81b6142d40c\") " pod="openshift-marketplace/community-operators-fbmpt" Oct 07 14:56:18 crc kubenswrapper[4959]: I1007 14:56:18.594961 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmncz\" (UniqueName: \"kubernetes.io/projected/71912490-8062-40d6-b950-e81b6142d40c-kube-api-access-hmncz\") pod \"community-operators-fbmpt\" (UID: \"71912490-8062-40d6-b950-e81b6142d40c\") " pod="openshift-marketplace/community-operators-fbmpt" Oct 07 14:56:18 crc kubenswrapper[4959]: I1007 14:56:18.595172 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71912490-8062-40d6-b950-e81b6142d40c-utilities\") pod \"community-operators-fbmpt\" (UID: \"71912490-8062-40d6-b950-e81b6142d40c\") " pod="openshift-marketplace/community-operators-fbmpt" Oct 07 14:56:18 crc kubenswrapper[4959]: I1007 14:56:18.595242 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71912490-8062-40d6-b950-e81b6142d40c-catalog-content\") pod \"community-operators-fbmpt\" (UID: \"71912490-8062-40d6-b950-e81b6142d40c\") " pod="openshift-marketplace/community-operators-fbmpt" Oct 07 14:56:18 crc kubenswrapper[4959]: I1007 14:56:18.595871 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71912490-8062-40d6-b950-e81b6142d40c-utilities\") pod \"community-operators-fbmpt\" (UID: \"71912490-8062-40d6-b950-e81b6142d40c\") " pod="openshift-marketplace/community-operators-fbmpt" Oct 07 14:56:18 crc kubenswrapper[4959]: I1007 14:56:18.595917 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71912490-8062-40d6-b950-e81b6142d40c-catalog-content\") pod \"community-operators-fbmpt\" (UID: \"71912490-8062-40d6-b950-e81b6142d40c\") " pod="openshift-marketplace/community-operators-fbmpt" Oct 07 14:56:18 crc kubenswrapper[4959]: I1007 14:56:18.624167 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmncz\" (UniqueName: \"kubernetes.io/projected/71912490-8062-40d6-b950-e81b6142d40c-kube-api-access-hmncz\") pod \"community-operators-fbmpt\" (UID: \"71912490-8062-40d6-b950-e81b6142d40c\") " pod="openshift-marketplace/community-operators-fbmpt" Oct 07 14:56:18 crc kubenswrapper[4959]: I1007 14:56:18.773497 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fbmpt" Oct 07 14:56:19 crc kubenswrapper[4959]: I1007 14:56:19.342588 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fbmpt"] Oct 07 14:56:19 crc kubenswrapper[4959]: I1007 14:56:19.435075 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fbmpt" event={"ID":"71912490-8062-40d6-b950-e81b6142d40c","Type":"ContainerStarted","Data":"e543005ed930685bb8d112ae17e453c053f3cdf1384a7b348597a77ce8a8ce36"} Oct 07 14:56:20 crc kubenswrapper[4959]: I1007 14:56:20.448061 4959 generic.go:334] "Generic (PLEG): container finished" podID="71912490-8062-40d6-b950-e81b6142d40c" containerID="cfec75d86c9b48655caa5411121c711729c1f94fb45f389a203f0f8b0ccb4ec1" exitCode=0 Oct 07 14:56:20 crc kubenswrapper[4959]: I1007 14:56:20.448207 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fbmpt" event={"ID":"71912490-8062-40d6-b950-e81b6142d40c","Type":"ContainerDied","Data":"cfec75d86c9b48655caa5411121c711729c1f94fb45f389a203f0f8b0ccb4ec1"} Oct 07 14:56:20 crc kubenswrapper[4959]: I1007 14:56:20.630400 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:56:20 crc kubenswrapper[4959]: I1007 14:56:20.630778 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:56:21 crc kubenswrapper[4959]: I1007 14:56:21.460821 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fbmpt" event={"ID":"71912490-8062-40d6-b950-e81b6142d40c","Type":"ContainerStarted","Data":"a1eceac266084f6f1456585c847e227bea3126545aeb87752a48046cc898eb32"} Oct 07 14:56:22 crc kubenswrapper[4959]: I1007 14:56:22.472502 4959 generic.go:334] "Generic (PLEG): container finished" podID="71912490-8062-40d6-b950-e81b6142d40c" containerID="a1eceac266084f6f1456585c847e227bea3126545aeb87752a48046cc898eb32" exitCode=0 Oct 07 14:56:22 crc kubenswrapper[4959]: I1007 14:56:22.472692 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fbmpt" event={"ID":"71912490-8062-40d6-b950-e81b6142d40c","Type":"ContainerDied","Data":"a1eceac266084f6f1456585c847e227bea3126545aeb87752a48046cc898eb32"} Oct 07 14:56:24 crc kubenswrapper[4959]: I1007 14:56:24.518401 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fbmpt" event={"ID":"71912490-8062-40d6-b950-e81b6142d40c","Type":"ContainerStarted","Data":"3969f730eb5d6f7779c891aa7583e160bfb179dba7831718b2a25df6632c7862"} Oct 07 14:56:24 crc kubenswrapper[4959]: I1007 14:56:24.542852 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fbmpt" podStartSLOduration=4.113672323 podStartE2EDuration="6.542826591s" podCreationTimestamp="2025-10-07 14:56:18 +0000 UTC" firstStartedPulling="2025-10-07 14:56:20.450666231 +0000 UTC m=+4262.534070556" lastFinishedPulling="2025-10-07 14:56:22.879820499 +0000 UTC m=+4264.963224824" observedRunningTime="2025-10-07 14:56:24.539467695 +0000 UTC m=+4266.622872030" watchObservedRunningTime="2025-10-07 14:56:24.542826591 +0000 UTC m=+4266.626230946" Oct 07 14:56:28 crc kubenswrapper[4959]: I1007 14:56:28.775037 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fbmpt" Oct 07 14:56:28 crc kubenswrapper[4959]: I1007 14:56:28.776532 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fbmpt" Oct 07 14:56:28 crc kubenswrapper[4959]: I1007 14:56:28.839847 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fbmpt" Oct 07 14:56:29 crc kubenswrapper[4959]: I1007 14:56:29.621827 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fbmpt" Oct 07 14:56:29 crc kubenswrapper[4959]: I1007 14:56:29.682190 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fbmpt"] Oct 07 14:56:31 crc kubenswrapper[4959]: I1007 14:56:31.586889 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fbmpt" podUID="71912490-8062-40d6-b950-e81b6142d40c" containerName="registry-server" containerID="cri-o://3969f730eb5d6f7779c891aa7583e160bfb179dba7831718b2a25df6632c7862" gracePeriod=2 Oct 07 14:56:32 crc kubenswrapper[4959]: I1007 14:56:32.271995 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fbmpt" Oct 07 14:56:32 crc kubenswrapper[4959]: I1007 14:56:32.412481 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71912490-8062-40d6-b950-e81b6142d40c-catalog-content\") pod \"71912490-8062-40d6-b950-e81b6142d40c\" (UID: \"71912490-8062-40d6-b950-e81b6142d40c\") " Oct 07 14:56:32 crc kubenswrapper[4959]: I1007 14:56:32.412575 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmncz\" (UniqueName: \"kubernetes.io/projected/71912490-8062-40d6-b950-e81b6142d40c-kube-api-access-hmncz\") pod \"71912490-8062-40d6-b950-e81b6142d40c\" (UID: \"71912490-8062-40d6-b950-e81b6142d40c\") " Oct 07 14:56:32 crc kubenswrapper[4959]: I1007 14:56:32.412804 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71912490-8062-40d6-b950-e81b6142d40c-utilities\") pod \"71912490-8062-40d6-b950-e81b6142d40c\" (UID: \"71912490-8062-40d6-b950-e81b6142d40c\") " Oct 07 14:56:32 crc kubenswrapper[4959]: I1007 14:56:32.413998 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71912490-8062-40d6-b950-e81b6142d40c-utilities" (OuterVolumeSpecName: "utilities") pod "71912490-8062-40d6-b950-e81b6142d40c" (UID: "71912490-8062-40d6-b950-e81b6142d40c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:56:32 crc kubenswrapper[4959]: I1007 14:56:32.422026 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71912490-8062-40d6-b950-e81b6142d40c-kube-api-access-hmncz" (OuterVolumeSpecName: "kube-api-access-hmncz") pod "71912490-8062-40d6-b950-e81b6142d40c" (UID: "71912490-8062-40d6-b950-e81b6142d40c"). InnerVolumeSpecName "kube-api-access-hmncz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:56:32 crc kubenswrapper[4959]: I1007 14:56:32.471015 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71912490-8062-40d6-b950-e81b6142d40c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "71912490-8062-40d6-b950-e81b6142d40c" (UID: "71912490-8062-40d6-b950-e81b6142d40c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:56:32 crc kubenswrapper[4959]: I1007 14:56:32.515612 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71912490-8062-40d6-b950-e81b6142d40c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 14:56:32 crc kubenswrapper[4959]: I1007 14:56:32.515664 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmncz\" (UniqueName: \"kubernetes.io/projected/71912490-8062-40d6-b950-e81b6142d40c-kube-api-access-hmncz\") on node \"crc\" DevicePath \"\"" Oct 07 14:56:32 crc kubenswrapper[4959]: I1007 14:56:32.515678 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71912490-8062-40d6-b950-e81b6142d40c-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 14:56:32 crc kubenswrapper[4959]: I1007 14:56:32.602818 4959 generic.go:334] "Generic (PLEG): container finished" podID="71912490-8062-40d6-b950-e81b6142d40c" containerID="3969f730eb5d6f7779c891aa7583e160bfb179dba7831718b2a25df6632c7862" exitCode=0 Oct 07 14:56:32 crc kubenswrapper[4959]: I1007 14:56:32.602875 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fbmpt" event={"ID":"71912490-8062-40d6-b950-e81b6142d40c","Type":"ContainerDied","Data":"3969f730eb5d6f7779c891aa7583e160bfb179dba7831718b2a25df6632c7862"} Oct 07 14:56:32 crc kubenswrapper[4959]: I1007 14:56:32.602915 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fbmpt" event={"ID":"71912490-8062-40d6-b950-e81b6142d40c","Type":"ContainerDied","Data":"e543005ed930685bb8d112ae17e453c053f3cdf1384a7b348597a77ce8a8ce36"} Oct 07 14:56:32 crc kubenswrapper[4959]: I1007 14:56:32.602939 4959 scope.go:117] "RemoveContainer" containerID="3969f730eb5d6f7779c891aa7583e160bfb179dba7831718b2a25df6632c7862" Oct 07 14:56:32 crc kubenswrapper[4959]: I1007 14:56:32.602947 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fbmpt" Oct 07 14:56:32 crc kubenswrapper[4959]: I1007 14:56:32.636649 4959 scope.go:117] "RemoveContainer" containerID="a1eceac266084f6f1456585c847e227bea3126545aeb87752a48046cc898eb32" Oct 07 14:56:32 crc kubenswrapper[4959]: I1007 14:56:32.639889 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fbmpt"] Oct 07 14:56:32 crc kubenswrapper[4959]: I1007 14:56:32.651444 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fbmpt"] Oct 07 14:56:32 crc kubenswrapper[4959]: I1007 14:56:32.669070 4959 scope.go:117] "RemoveContainer" containerID="cfec75d86c9b48655caa5411121c711729c1f94fb45f389a203f0f8b0ccb4ec1" Oct 07 14:56:32 crc kubenswrapper[4959]: I1007 14:56:32.675454 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71912490-8062-40d6-b950-e81b6142d40c" path="/var/lib/kubelet/pods/71912490-8062-40d6-b950-e81b6142d40c/volumes" Oct 07 14:56:32 crc kubenswrapper[4959]: I1007 14:56:32.726885 4959 scope.go:117] "RemoveContainer" containerID="3969f730eb5d6f7779c891aa7583e160bfb179dba7831718b2a25df6632c7862" Oct 07 14:56:32 crc kubenswrapper[4959]: E1007 14:56:32.727479 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3969f730eb5d6f7779c891aa7583e160bfb179dba7831718b2a25df6632c7862\": container with ID starting with 3969f730eb5d6f7779c891aa7583e160bfb179dba7831718b2a25df6632c7862 not found: ID does not exist" containerID="3969f730eb5d6f7779c891aa7583e160bfb179dba7831718b2a25df6632c7862" Oct 07 14:56:32 crc kubenswrapper[4959]: I1007 14:56:32.727522 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3969f730eb5d6f7779c891aa7583e160bfb179dba7831718b2a25df6632c7862"} err="failed to get container status \"3969f730eb5d6f7779c891aa7583e160bfb179dba7831718b2a25df6632c7862\": rpc error: code = NotFound desc = could not find container \"3969f730eb5d6f7779c891aa7583e160bfb179dba7831718b2a25df6632c7862\": container with ID starting with 3969f730eb5d6f7779c891aa7583e160bfb179dba7831718b2a25df6632c7862 not found: ID does not exist" Oct 07 14:56:32 crc kubenswrapper[4959]: I1007 14:56:32.727554 4959 scope.go:117] "RemoveContainer" containerID="a1eceac266084f6f1456585c847e227bea3126545aeb87752a48046cc898eb32" Oct 07 14:56:32 crc kubenswrapper[4959]: E1007 14:56:32.728125 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1eceac266084f6f1456585c847e227bea3126545aeb87752a48046cc898eb32\": container with ID starting with a1eceac266084f6f1456585c847e227bea3126545aeb87752a48046cc898eb32 not found: ID does not exist" containerID="a1eceac266084f6f1456585c847e227bea3126545aeb87752a48046cc898eb32" Oct 07 14:56:32 crc kubenswrapper[4959]: I1007 14:56:32.728186 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1eceac266084f6f1456585c847e227bea3126545aeb87752a48046cc898eb32"} err="failed to get container status \"a1eceac266084f6f1456585c847e227bea3126545aeb87752a48046cc898eb32\": rpc error: code = NotFound desc = could not find container \"a1eceac266084f6f1456585c847e227bea3126545aeb87752a48046cc898eb32\": container with ID starting with a1eceac266084f6f1456585c847e227bea3126545aeb87752a48046cc898eb32 not found: ID does not exist" Oct 07 14:56:32 crc kubenswrapper[4959]: I1007 14:56:32.728230 4959 scope.go:117] "RemoveContainer" containerID="cfec75d86c9b48655caa5411121c711729c1f94fb45f389a203f0f8b0ccb4ec1" Oct 07 14:56:32 crc kubenswrapper[4959]: E1007 14:56:32.728896 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfec75d86c9b48655caa5411121c711729c1f94fb45f389a203f0f8b0ccb4ec1\": container with ID starting with cfec75d86c9b48655caa5411121c711729c1f94fb45f389a203f0f8b0ccb4ec1 not found: ID does not exist" containerID="cfec75d86c9b48655caa5411121c711729c1f94fb45f389a203f0f8b0ccb4ec1" Oct 07 14:56:32 crc kubenswrapper[4959]: I1007 14:56:32.728975 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfec75d86c9b48655caa5411121c711729c1f94fb45f389a203f0f8b0ccb4ec1"} err="failed to get container status \"cfec75d86c9b48655caa5411121c711729c1f94fb45f389a203f0f8b0ccb4ec1\": rpc error: code = NotFound desc = could not find container \"cfec75d86c9b48655caa5411121c711729c1f94fb45f389a203f0f8b0ccb4ec1\": container with ID starting with cfec75d86c9b48655caa5411121c711729c1f94fb45f389a203f0f8b0ccb4ec1 not found: ID does not exist" Oct 07 14:56:50 crc kubenswrapper[4959]: I1007 14:56:50.630972 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:56:50 crc kubenswrapper[4959]: I1007 14:56:50.632013 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:57:20 crc kubenswrapper[4959]: I1007 14:57:20.629978 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 14:57:20 crc kubenswrapper[4959]: I1007 14:57:20.630729 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 14:57:20 crc kubenswrapper[4959]: I1007 14:57:20.630799 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 14:57:20 crc kubenswrapper[4959]: I1007 14:57:20.631662 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204"} pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 14:57:20 crc kubenswrapper[4959]: I1007 14:57:20.631725 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" containerID="cri-o://9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204" gracePeriod=600 Oct 07 14:57:20 crc kubenswrapper[4959]: E1007 14:57:20.772460 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:57:21 crc kubenswrapper[4959]: I1007 14:57:21.088714 4959 generic.go:334] "Generic (PLEG): container finished" podID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerID="9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204" exitCode=0 Oct 07 14:57:21 crc kubenswrapper[4959]: I1007 14:57:21.088811 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerDied","Data":"9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204"} Oct 07 14:57:21 crc kubenswrapper[4959]: I1007 14:57:21.089140 4959 scope.go:117] "RemoveContainer" containerID="4d93ebe2980eadde9513ef305d1bfacd8c628c9df14fa9718e12e94588e90557" Oct 07 14:57:21 crc kubenswrapper[4959]: I1007 14:57:21.090392 4959 scope.go:117] "RemoveContainer" containerID="9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204" Oct 07 14:57:21 crc kubenswrapper[4959]: E1007 14:57:21.090947 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:57:35 crc kubenswrapper[4959]: I1007 14:57:35.654298 4959 scope.go:117] "RemoveContainer" containerID="9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204" Oct 07 14:57:35 crc kubenswrapper[4959]: E1007 14:57:35.655573 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:57:48 crc kubenswrapper[4959]: I1007 14:57:48.661860 4959 scope.go:117] "RemoveContainer" containerID="9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204" Oct 07 14:57:48 crc kubenswrapper[4959]: E1007 14:57:48.662697 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:58:03 crc kubenswrapper[4959]: I1007 14:58:03.654868 4959 scope.go:117] "RemoveContainer" containerID="9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204" Oct 07 14:58:03 crc kubenswrapper[4959]: E1007 14:58:03.656431 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:58:18 crc kubenswrapper[4959]: I1007 14:58:18.659567 4959 scope.go:117] "RemoveContainer" containerID="9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204" Oct 07 14:58:18 crc kubenswrapper[4959]: E1007 14:58:18.660379 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:58:29 crc kubenswrapper[4959]: I1007 14:58:29.654908 4959 scope.go:117] "RemoveContainer" containerID="9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204" Oct 07 14:58:29 crc kubenswrapper[4959]: E1007 14:58:29.656171 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:58:42 crc kubenswrapper[4959]: I1007 14:58:42.653502 4959 scope.go:117] "RemoveContainer" containerID="9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204" Oct 07 14:58:42 crc kubenswrapper[4959]: E1007 14:58:42.654302 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:58:43 crc kubenswrapper[4959]: I1007 14:58:43.051029 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dl6rh"] Oct 07 14:58:43 crc kubenswrapper[4959]: E1007 14:58:43.052011 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71912490-8062-40d6-b950-e81b6142d40c" containerName="extract-utilities" Oct 07 14:58:43 crc kubenswrapper[4959]: I1007 14:58:43.052042 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="71912490-8062-40d6-b950-e81b6142d40c" containerName="extract-utilities" Oct 07 14:58:43 crc kubenswrapper[4959]: E1007 14:58:43.052059 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71912490-8062-40d6-b950-e81b6142d40c" containerName="registry-server" Oct 07 14:58:43 crc kubenswrapper[4959]: I1007 14:58:43.052069 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="71912490-8062-40d6-b950-e81b6142d40c" containerName="registry-server" Oct 07 14:58:43 crc kubenswrapper[4959]: E1007 14:58:43.052093 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71912490-8062-40d6-b950-e81b6142d40c" containerName="extract-content" Oct 07 14:58:43 crc kubenswrapper[4959]: I1007 14:58:43.052124 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="71912490-8062-40d6-b950-e81b6142d40c" containerName="extract-content" Oct 07 14:58:43 crc kubenswrapper[4959]: I1007 14:58:43.052366 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="71912490-8062-40d6-b950-e81b6142d40c" containerName="registry-server" Oct 07 14:58:43 crc kubenswrapper[4959]: I1007 14:58:43.054224 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dl6rh" Oct 07 14:58:43 crc kubenswrapper[4959]: I1007 14:58:43.063758 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dl6rh"] Oct 07 14:58:43 crc kubenswrapper[4959]: I1007 14:58:43.075626 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llzfn\" (UniqueName: \"kubernetes.io/projected/0a86d3f1-31eb-4e2e-acdd-592802599d1b-kube-api-access-llzfn\") pod \"certified-operators-dl6rh\" (UID: \"0a86d3f1-31eb-4e2e-acdd-592802599d1b\") " pod="openshift-marketplace/certified-operators-dl6rh" Oct 07 14:58:43 crc kubenswrapper[4959]: I1007 14:58:43.076157 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a86d3f1-31eb-4e2e-acdd-592802599d1b-catalog-content\") pod \"certified-operators-dl6rh\" (UID: \"0a86d3f1-31eb-4e2e-acdd-592802599d1b\") " pod="openshift-marketplace/certified-operators-dl6rh" Oct 07 14:58:43 crc kubenswrapper[4959]: I1007 14:58:43.076306 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a86d3f1-31eb-4e2e-acdd-592802599d1b-utilities\") pod \"certified-operators-dl6rh\" (UID: \"0a86d3f1-31eb-4e2e-acdd-592802599d1b\") " pod="openshift-marketplace/certified-operators-dl6rh" Oct 07 14:58:43 crc kubenswrapper[4959]: I1007 14:58:43.180866 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a86d3f1-31eb-4e2e-acdd-592802599d1b-utilities\") pod \"certified-operators-dl6rh\" (UID: \"0a86d3f1-31eb-4e2e-acdd-592802599d1b\") " pod="openshift-marketplace/certified-operators-dl6rh" Oct 07 14:58:43 crc kubenswrapper[4959]: I1007 14:58:43.181037 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llzfn\" (UniqueName: \"kubernetes.io/projected/0a86d3f1-31eb-4e2e-acdd-592802599d1b-kube-api-access-llzfn\") pod \"certified-operators-dl6rh\" (UID: \"0a86d3f1-31eb-4e2e-acdd-592802599d1b\") " pod="openshift-marketplace/certified-operators-dl6rh" Oct 07 14:58:43 crc kubenswrapper[4959]: I1007 14:58:43.181100 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a86d3f1-31eb-4e2e-acdd-592802599d1b-catalog-content\") pod \"certified-operators-dl6rh\" (UID: \"0a86d3f1-31eb-4e2e-acdd-592802599d1b\") " pod="openshift-marketplace/certified-operators-dl6rh" Oct 07 14:58:43 crc kubenswrapper[4959]: I1007 14:58:43.181715 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a86d3f1-31eb-4e2e-acdd-592802599d1b-utilities\") pod \"certified-operators-dl6rh\" (UID: \"0a86d3f1-31eb-4e2e-acdd-592802599d1b\") " pod="openshift-marketplace/certified-operators-dl6rh" Oct 07 14:58:43 crc kubenswrapper[4959]: I1007 14:58:43.181770 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a86d3f1-31eb-4e2e-acdd-592802599d1b-catalog-content\") pod \"certified-operators-dl6rh\" (UID: \"0a86d3f1-31eb-4e2e-acdd-592802599d1b\") " pod="openshift-marketplace/certified-operators-dl6rh" Oct 07 14:58:43 crc kubenswrapper[4959]: I1007 14:58:43.210861 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llzfn\" (UniqueName: \"kubernetes.io/projected/0a86d3f1-31eb-4e2e-acdd-592802599d1b-kube-api-access-llzfn\") pod \"certified-operators-dl6rh\" (UID: \"0a86d3f1-31eb-4e2e-acdd-592802599d1b\") " pod="openshift-marketplace/certified-operators-dl6rh" Oct 07 14:58:43 crc kubenswrapper[4959]: I1007 14:58:43.378050 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dl6rh" Oct 07 14:58:43 crc kubenswrapper[4959]: I1007 14:58:43.994977 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dl6rh"] Oct 07 14:58:44 crc kubenswrapper[4959]: I1007 14:58:44.963774 4959 generic.go:334] "Generic (PLEG): container finished" podID="0a86d3f1-31eb-4e2e-acdd-592802599d1b" containerID="db95870396910987968e785a80868a21e6e391aabff49ef075be36ceb1a29612" exitCode=0 Oct 07 14:58:44 crc kubenswrapper[4959]: I1007 14:58:44.964356 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dl6rh" event={"ID":"0a86d3f1-31eb-4e2e-acdd-592802599d1b","Type":"ContainerDied","Data":"db95870396910987968e785a80868a21e6e391aabff49ef075be36ceb1a29612"} Oct 07 14:58:44 crc kubenswrapper[4959]: I1007 14:58:44.965343 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dl6rh" event={"ID":"0a86d3f1-31eb-4e2e-acdd-592802599d1b","Type":"ContainerStarted","Data":"e86e608debe8f259593439578668861d60a8f8bbdd85a64951435f440cabed7e"} Oct 07 14:58:51 crc kubenswrapper[4959]: I1007 14:58:51.036911 4959 generic.go:334] "Generic (PLEG): container finished" podID="0a86d3f1-31eb-4e2e-acdd-592802599d1b" containerID="53900151a28a4826938061e6e414780b0d86a5be767310b97e6d8578737f06f5" exitCode=0 Oct 07 14:58:51 crc kubenswrapper[4959]: I1007 14:58:51.037375 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dl6rh" event={"ID":"0a86d3f1-31eb-4e2e-acdd-592802599d1b","Type":"ContainerDied","Data":"53900151a28a4826938061e6e414780b0d86a5be767310b97e6d8578737f06f5"} Oct 07 14:58:53 crc kubenswrapper[4959]: I1007 14:58:53.060671 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dl6rh" event={"ID":"0a86d3f1-31eb-4e2e-acdd-592802599d1b","Type":"ContainerStarted","Data":"b863572d8c2fe286fe00acea31d5fa8b8d329098c62ab0da3eaf4ee20af82cee"} Oct 07 14:58:53 crc kubenswrapper[4959]: I1007 14:58:53.083694 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dl6rh" podStartSLOduration=3.538308906 podStartE2EDuration="10.083672757s" podCreationTimestamp="2025-10-07 14:58:43 +0000 UTC" firstStartedPulling="2025-10-07 14:58:44.965924922 +0000 UTC m=+4407.049329247" lastFinishedPulling="2025-10-07 14:58:51.511288773 +0000 UTC m=+4413.594693098" observedRunningTime="2025-10-07 14:58:53.083194985 +0000 UTC m=+4415.166599320" watchObservedRunningTime="2025-10-07 14:58:53.083672757 +0000 UTC m=+4415.167077082" Oct 07 14:58:53 crc kubenswrapper[4959]: I1007 14:58:53.379139 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dl6rh" Oct 07 14:58:53 crc kubenswrapper[4959]: I1007 14:58:53.379192 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dl6rh" Oct 07 14:58:53 crc kubenswrapper[4959]: I1007 14:58:53.654169 4959 scope.go:117] "RemoveContainer" containerID="9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204" Oct 07 14:58:53 crc kubenswrapper[4959]: E1007 14:58:53.654595 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:58:54 crc kubenswrapper[4959]: I1007 14:58:54.427091 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-dl6rh" podUID="0a86d3f1-31eb-4e2e-acdd-592802599d1b" containerName="registry-server" probeResult="failure" output=< Oct 07 14:58:54 crc kubenswrapper[4959]: timeout: failed to connect service ":50051" within 1s Oct 07 14:58:54 crc kubenswrapper[4959]: > Oct 07 14:59:03 crc kubenswrapper[4959]: I1007 14:59:03.430860 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dl6rh" Oct 07 14:59:03 crc kubenswrapper[4959]: I1007 14:59:03.490215 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dl6rh" Oct 07 14:59:03 crc kubenswrapper[4959]: I1007 14:59:03.568208 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dl6rh"] Oct 07 14:59:03 crc kubenswrapper[4959]: I1007 14:59:03.689597 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f6t5x"] Oct 07 14:59:03 crc kubenswrapper[4959]: I1007 14:59:03.690022 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-f6t5x" podUID="f1a1cb02-d50a-441e-8ec6-0d2a8a12b255" containerName="registry-server" containerID="cri-o://b84a3561fec6eafb45bb30accb651de9437d81e3664414cd5b137f3a4a53d628" gracePeriod=2 Oct 07 14:59:04 crc kubenswrapper[4959]: I1007 14:59:04.171377 4959 generic.go:334] "Generic (PLEG): container finished" podID="f1a1cb02-d50a-441e-8ec6-0d2a8a12b255" containerID="b84a3561fec6eafb45bb30accb651de9437d81e3664414cd5b137f3a4a53d628" exitCode=0 Oct 07 14:59:04 crc kubenswrapper[4959]: I1007 14:59:04.171468 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f6t5x" event={"ID":"f1a1cb02-d50a-441e-8ec6-0d2a8a12b255","Type":"ContainerDied","Data":"b84a3561fec6eafb45bb30accb651de9437d81e3664414cd5b137f3a4a53d628"} Oct 07 14:59:04 crc kubenswrapper[4959]: I1007 14:59:04.807367 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f6t5x" Oct 07 14:59:04 crc kubenswrapper[4959]: I1007 14:59:04.907769 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sln78\" (UniqueName: \"kubernetes.io/projected/f1a1cb02-d50a-441e-8ec6-0d2a8a12b255-kube-api-access-sln78\") pod \"f1a1cb02-d50a-441e-8ec6-0d2a8a12b255\" (UID: \"f1a1cb02-d50a-441e-8ec6-0d2a8a12b255\") " Oct 07 14:59:04 crc kubenswrapper[4959]: I1007 14:59:04.907977 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1a1cb02-d50a-441e-8ec6-0d2a8a12b255-utilities\") pod \"f1a1cb02-d50a-441e-8ec6-0d2a8a12b255\" (UID: \"f1a1cb02-d50a-441e-8ec6-0d2a8a12b255\") " Oct 07 14:59:04 crc kubenswrapper[4959]: I1007 14:59:04.908082 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1a1cb02-d50a-441e-8ec6-0d2a8a12b255-catalog-content\") pod \"f1a1cb02-d50a-441e-8ec6-0d2a8a12b255\" (UID: \"f1a1cb02-d50a-441e-8ec6-0d2a8a12b255\") " Oct 07 14:59:05 crc kubenswrapper[4959]: I1007 14:59:05.019233 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1a1cb02-d50a-441e-8ec6-0d2a8a12b255-utilities" (OuterVolumeSpecName: "utilities") pod "f1a1cb02-d50a-441e-8ec6-0d2a8a12b255" (UID: "f1a1cb02-d50a-441e-8ec6-0d2a8a12b255"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:59:05 crc kubenswrapper[4959]: I1007 14:59:05.087768 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1a1cb02-d50a-441e-8ec6-0d2a8a12b255-kube-api-access-sln78" (OuterVolumeSpecName: "kube-api-access-sln78") pod "f1a1cb02-d50a-441e-8ec6-0d2a8a12b255" (UID: "f1a1cb02-d50a-441e-8ec6-0d2a8a12b255"). InnerVolumeSpecName "kube-api-access-sln78". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 14:59:05 crc kubenswrapper[4959]: I1007 14:59:05.114803 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sln78\" (UniqueName: \"kubernetes.io/projected/f1a1cb02-d50a-441e-8ec6-0d2a8a12b255-kube-api-access-sln78\") on node \"crc\" DevicePath \"\"" Oct 07 14:59:05 crc kubenswrapper[4959]: I1007 14:59:05.114843 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1a1cb02-d50a-441e-8ec6-0d2a8a12b255-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 14:59:05 crc kubenswrapper[4959]: I1007 14:59:05.187008 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f6t5x" event={"ID":"f1a1cb02-d50a-441e-8ec6-0d2a8a12b255","Type":"ContainerDied","Data":"857caa629fa34aed0e8d8c626600ece6fdf5e35fe9520d0fa20aa8cd41d0e7df"} Oct 07 14:59:05 crc kubenswrapper[4959]: I1007 14:59:05.187378 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f6t5x" Oct 07 14:59:05 crc kubenswrapper[4959]: I1007 14:59:05.187401 4959 scope.go:117] "RemoveContainer" containerID="b84a3561fec6eafb45bb30accb651de9437d81e3664414cd5b137f3a4a53d628" Oct 07 14:59:05 crc kubenswrapper[4959]: I1007 14:59:05.191299 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1a1cb02-d50a-441e-8ec6-0d2a8a12b255-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f1a1cb02-d50a-441e-8ec6-0d2a8a12b255" (UID: "f1a1cb02-d50a-441e-8ec6-0d2a8a12b255"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 14:59:05 crc kubenswrapper[4959]: I1007 14:59:05.219260 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1a1cb02-d50a-441e-8ec6-0d2a8a12b255-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 14:59:05 crc kubenswrapper[4959]: I1007 14:59:05.224266 4959 scope.go:117] "RemoveContainer" containerID="3ee358ca133d56ccf7ea430882408f812f1b3e4154fe4396ba81cd80bdf651e5" Oct 07 14:59:05 crc kubenswrapper[4959]: I1007 14:59:05.293438 4959 scope.go:117] "RemoveContainer" containerID="51a81b5769f162334215d9101493fa7a61f0f787fa285594d20aa0c32f9b4979" Oct 07 14:59:05 crc kubenswrapper[4959]: I1007 14:59:05.521069 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f6t5x"] Oct 07 14:59:05 crc kubenswrapper[4959]: I1007 14:59:05.532473 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-f6t5x"] Oct 07 14:59:05 crc kubenswrapper[4959]: I1007 14:59:05.654358 4959 scope.go:117] "RemoveContainer" containerID="9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204" Oct 07 14:59:05 crc kubenswrapper[4959]: E1007 14:59:05.654774 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:59:06 crc kubenswrapper[4959]: I1007 14:59:06.665260 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1a1cb02-d50a-441e-8ec6-0d2a8a12b255" path="/var/lib/kubelet/pods/f1a1cb02-d50a-441e-8ec6-0d2a8a12b255/volumes" Oct 07 14:59:20 crc kubenswrapper[4959]: I1007 14:59:20.654166 4959 scope.go:117] "RemoveContainer" containerID="9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204" Oct 07 14:59:20 crc kubenswrapper[4959]: E1007 14:59:20.654988 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:59:31 crc kubenswrapper[4959]: I1007 14:59:31.654051 4959 scope.go:117] "RemoveContainer" containerID="9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204" Oct 07 14:59:31 crc kubenswrapper[4959]: E1007 14:59:31.654854 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:59:42 crc kubenswrapper[4959]: I1007 14:59:42.654673 4959 scope.go:117] "RemoveContainer" containerID="9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204" Oct 07 14:59:42 crc kubenswrapper[4959]: E1007 14:59:42.655423 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 14:59:53 crc kubenswrapper[4959]: I1007 14:59:53.653277 4959 scope.go:117] "RemoveContainer" containerID="9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204" Oct 07 14:59:53 crc kubenswrapper[4959]: E1007 14:59:53.653982 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:00:00 crc kubenswrapper[4959]: I1007 15:00:00.143543 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330820-j5lrc"] Oct 07 15:00:00 crc kubenswrapper[4959]: E1007 15:00:00.144458 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1a1cb02-d50a-441e-8ec6-0d2a8a12b255" containerName="extract-content" Oct 07 15:00:00 crc kubenswrapper[4959]: I1007 15:00:00.144471 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1a1cb02-d50a-441e-8ec6-0d2a8a12b255" containerName="extract-content" Oct 07 15:00:00 crc kubenswrapper[4959]: E1007 15:00:00.144485 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1a1cb02-d50a-441e-8ec6-0d2a8a12b255" containerName="extract-utilities" Oct 07 15:00:00 crc kubenswrapper[4959]: I1007 15:00:00.144492 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1a1cb02-d50a-441e-8ec6-0d2a8a12b255" containerName="extract-utilities" Oct 07 15:00:00 crc kubenswrapper[4959]: E1007 15:00:00.144518 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1a1cb02-d50a-441e-8ec6-0d2a8a12b255" containerName="registry-server" Oct 07 15:00:00 crc kubenswrapper[4959]: I1007 15:00:00.144525 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1a1cb02-d50a-441e-8ec6-0d2a8a12b255" containerName="registry-server" Oct 07 15:00:00 crc kubenswrapper[4959]: I1007 15:00:00.144710 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1a1cb02-d50a-441e-8ec6-0d2a8a12b255" containerName="registry-server" Oct 07 15:00:00 crc kubenswrapper[4959]: I1007 15:00:00.145650 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330820-j5lrc" Oct 07 15:00:00 crc kubenswrapper[4959]: I1007 15:00:00.147646 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 07 15:00:00 crc kubenswrapper[4959]: I1007 15:00:00.147713 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 07 15:00:00 crc kubenswrapper[4959]: I1007 15:00:00.158389 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330820-j5lrc"] Oct 07 15:00:00 crc kubenswrapper[4959]: I1007 15:00:00.306774 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/63f04186-4fb9-4518-b19e-b0a7bf573df0-config-volume\") pod \"collect-profiles-29330820-j5lrc\" (UID: \"63f04186-4fb9-4518-b19e-b0a7bf573df0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330820-j5lrc" Oct 07 15:00:00 crc kubenswrapper[4959]: I1007 15:00:00.306848 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/63f04186-4fb9-4518-b19e-b0a7bf573df0-secret-volume\") pod \"collect-profiles-29330820-j5lrc\" (UID: \"63f04186-4fb9-4518-b19e-b0a7bf573df0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330820-j5lrc" Oct 07 15:00:00 crc kubenswrapper[4959]: I1007 15:00:00.307051 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67lhg\" (UniqueName: \"kubernetes.io/projected/63f04186-4fb9-4518-b19e-b0a7bf573df0-kube-api-access-67lhg\") pod \"collect-profiles-29330820-j5lrc\" (UID: \"63f04186-4fb9-4518-b19e-b0a7bf573df0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330820-j5lrc" Oct 07 15:00:00 crc kubenswrapper[4959]: I1007 15:00:00.409499 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/63f04186-4fb9-4518-b19e-b0a7bf573df0-secret-volume\") pod \"collect-profiles-29330820-j5lrc\" (UID: \"63f04186-4fb9-4518-b19e-b0a7bf573df0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330820-j5lrc" Oct 07 15:00:00 crc kubenswrapper[4959]: I1007 15:00:00.409561 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/63f04186-4fb9-4518-b19e-b0a7bf573df0-config-volume\") pod \"collect-profiles-29330820-j5lrc\" (UID: \"63f04186-4fb9-4518-b19e-b0a7bf573df0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330820-j5lrc" Oct 07 15:00:00 crc kubenswrapper[4959]: I1007 15:00:00.409728 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67lhg\" (UniqueName: \"kubernetes.io/projected/63f04186-4fb9-4518-b19e-b0a7bf573df0-kube-api-access-67lhg\") pod \"collect-profiles-29330820-j5lrc\" (UID: \"63f04186-4fb9-4518-b19e-b0a7bf573df0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330820-j5lrc" Oct 07 15:00:00 crc kubenswrapper[4959]: I1007 15:00:00.410729 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/63f04186-4fb9-4518-b19e-b0a7bf573df0-config-volume\") pod \"collect-profiles-29330820-j5lrc\" (UID: \"63f04186-4fb9-4518-b19e-b0a7bf573df0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330820-j5lrc" Oct 07 15:00:00 crc kubenswrapper[4959]: I1007 15:00:00.416734 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/63f04186-4fb9-4518-b19e-b0a7bf573df0-secret-volume\") pod \"collect-profiles-29330820-j5lrc\" (UID: \"63f04186-4fb9-4518-b19e-b0a7bf573df0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330820-j5lrc" Oct 07 15:00:00 crc kubenswrapper[4959]: I1007 15:00:00.428287 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67lhg\" (UniqueName: \"kubernetes.io/projected/63f04186-4fb9-4518-b19e-b0a7bf573df0-kube-api-access-67lhg\") pod \"collect-profiles-29330820-j5lrc\" (UID: \"63f04186-4fb9-4518-b19e-b0a7bf573df0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330820-j5lrc" Oct 07 15:00:00 crc kubenswrapper[4959]: I1007 15:00:00.482459 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330820-j5lrc" Oct 07 15:00:00 crc kubenswrapper[4959]: I1007 15:00:00.972996 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330820-j5lrc"] Oct 07 15:00:01 crc kubenswrapper[4959]: I1007 15:00:01.703972 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330820-j5lrc" event={"ID":"63f04186-4fb9-4518-b19e-b0a7bf573df0","Type":"ContainerStarted","Data":"049810a597be74998d30a03070d4f11a3c8ae4ec242bc4ede29616ab3dfc947a"} Oct 07 15:00:01 crc kubenswrapper[4959]: I1007 15:00:01.704692 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330820-j5lrc" event={"ID":"63f04186-4fb9-4518-b19e-b0a7bf573df0","Type":"ContainerStarted","Data":"facc7cd6739df576282db229690d473cfa1d905c6d30d2239b2adc477926ea24"} Oct 07 15:00:01 crc kubenswrapper[4959]: I1007 15:00:01.723575 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29330820-j5lrc" podStartSLOduration=1.7235415760000001 podStartE2EDuration="1.723541576s" podCreationTimestamp="2025-10-07 15:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 15:00:01.720274333 +0000 UTC m=+4483.803678648" watchObservedRunningTime="2025-10-07 15:00:01.723541576 +0000 UTC m=+4483.806945921" Oct 07 15:00:02 crc kubenswrapper[4959]: E1007 15:00:02.041905 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63f04186_4fb9_4518_b19e_b0a7bf573df0.slice/crio-049810a597be74998d30a03070d4f11a3c8ae4ec242bc4ede29616ab3dfc947a.scope\": RecentStats: unable to find data in memory cache]" Oct 07 15:00:02 crc kubenswrapper[4959]: I1007 15:00:02.715450 4959 generic.go:334] "Generic (PLEG): container finished" podID="63f04186-4fb9-4518-b19e-b0a7bf573df0" containerID="049810a597be74998d30a03070d4f11a3c8ae4ec242bc4ede29616ab3dfc947a" exitCode=0 Oct 07 15:00:02 crc kubenswrapper[4959]: I1007 15:00:02.715518 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330820-j5lrc" event={"ID":"63f04186-4fb9-4518-b19e-b0a7bf573df0","Type":"ContainerDied","Data":"049810a597be74998d30a03070d4f11a3c8ae4ec242bc4ede29616ab3dfc947a"} Oct 07 15:00:04 crc kubenswrapper[4959]: I1007 15:00:04.235171 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330820-j5lrc" Oct 07 15:00:04 crc kubenswrapper[4959]: I1007 15:00:04.390291 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67lhg\" (UniqueName: \"kubernetes.io/projected/63f04186-4fb9-4518-b19e-b0a7bf573df0-kube-api-access-67lhg\") pod \"63f04186-4fb9-4518-b19e-b0a7bf573df0\" (UID: \"63f04186-4fb9-4518-b19e-b0a7bf573df0\") " Oct 07 15:00:04 crc kubenswrapper[4959]: I1007 15:00:04.390649 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/63f04186-4fb9-4518-b19e-b0a7bf573df0-config-volume\") pod \"63f04186-4fb9-4518-b19e-b0a7bf573df0\" (UID: \"63f04186-4fb9-4518-b19e-b0a7bf573df0\") " Oct 07 15:00:04 crc kubenswrapper[4959]: I1007 15:00:04.390922 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/63f04186-4fb9-4518-b19e-b0a7bf573df0-secret-volume\") pod \"63f04186-4fb9-4518-b19e-b0a7bf573df0\" (UID: \"63f04186-4fb9-4518-b19e-b0a7bf573df0\") " Oct 07 15:00:04 crc kubenswrapper[4959]: I1007 15:00:04.391328 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63f04186-4fb9-4518-b19e-b0a7bf573df0-config-volume" (OuterVolumeSpecName: "config-volume") pod "63f04186-4fb9-4518-b19e-b0a7bf573df0" (UID: "63f04186-4fb9-4518-b19e-b0a7bf573df0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 15:00:04 crc kubenswrapper[4959]: I1007 15:00:04.391839 4959 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/63f04186-4fb9-4518-b19e-b0a7bf573df0-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 15:00:04 crc kubenswrapper[4959]: I1007 15:00:04.396627 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63f04186-4fb9-4518-b19e-b0a7bf573df0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "63f04186-4fb9-4518-b19e-b0a7bf573df0" (UID: "63f04186-4fb9-4518-b19e-b0a7bf573df0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:00:04 crc kubenswrapper[4959]: I1007 15:00:04.396897 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63f04186-4fb9-4518-b19e-b0a7bf573df0-kube-api-access-67lhg" (OuterVolumeSpecName: "kube-api-access-67lhg") pod "63f04186-4fb9-4518-b19e-b0a7bf573df0" (UID: "63f04186-4fb9-4518-b19e-b0a7bf573df0"). InnerVolumeSpecName "kube-api-access-67lhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 15:00:04 crc kubenswrapper[4959]: I1007 15:00:04.493791 4959 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/63f04186-4fb9-4518-b19e-b0a7bf573df0-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 07 15:00:04 crc kubenswrapper[4959]: I1007 15:00:04.493822 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67lhg\" (UniqueName: \"kubernetes.io/projected/63f04186-4fb9-4518-b19e-b0a7bf573df0-kube-api-access-67lhg\") on node \"crc\" DevicePath \"\"" Oct 07 15:00:04 crc kubenswrapper[4959]: I1007 15:00:04.735614 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330820-j5lrc" event={"ID":"63f04186-4fb9-4518-b19e-b0a7bf573df0","Type":"ContainerDied","Data":"facc7cd6739df576282db229690d473cfa1d905c6d30d2239b2adc477926ea24"} Oct 07 15:00:04 crc kubenswrapper[4959]: I1007 15:00:04.735674 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="facc7cd6739df576282db229690d473cfa1d905c6d30d2239b2adc477926ea24" Oct 07 15:00:04 crc kubenswrapper[4959]: I1007 15:00:04.736342 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330820-j5lrc" Oct 07 15:00:04 crc kubenswrapper[4959]: I1007 15:00:04.818952 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330775-flfr7"] Oct 07 15:00:04 crc kubenswrapper[4959]: I1007 15:00:04.829961 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330775-flfr7"] Oct 07 15:00:05 crc kubenswrapper[4959]: I1007 15:00:05.654241 4959 scope.go:117] "RemoveContainer" containerID="9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204" Oct 07 15:00:05 crc kubenswrapper[4959]: E1007 15:00:05.654993 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:00:06 crc kubenswrapper[4959]: I1007 15:00:06.664310 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba6d1d1d-89bf-4c7c-9266-cebc0505e53c" path="/var/lib/kubelet/pods/ba6d1d1d-89bf-4c7c-9266-cebc0505e53c/volumes" Oct 07 15:00:20 crc kubenswrapper[4959]: I1007 15:00:20.654779 4959 scope.go:117] "RemoveContainer" containerID="9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204" Oct 07 15:00:20 crc kubenswrapper[4959]: E1007 15:00:20.655888 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:00:31 crc kubenswrapper[4959]: I1007 15:00:31.654745 4959 scope.go:117] "RemoveContainer" containerID="9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204" Oct 07 15:00:31 crc kubenswrapper[4959]: E1007 15:00:31.656043 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:00:31 crc kubenswrapper[4959]: I1007 15:00:31.917391 4959 scope.go:117] "RemoveContainer" containerID="439abfcf60e7afaf0f1b838e06043c104500921fe892d2fc7c1013dd14ad616d" Oct 07 15:00:44 crc kubenswrapper[4959]: I1007 15:00:44.653812 4959 scope.go:117] "RemoveContainer" containerID="9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204" Oct 07 15:00:44 crc kubenswrapper[4959]: E1007 15:00:44.655428 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:00:55 crc kubenswrapper[4959]: I1007 15:00:55.654236 4959 scope.go:117] "RemoveContainer" containerID="9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204" Oct 07 15:00:55 crc kubenswrapper[4959]: E1007 15:00:55.655381 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:01:00 crc kubenswrapper[4959]: I1007 15:01:00.150903 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29330821-dpxgj"] Oct 07 15:01:00 crc kubenswrapper[4959]: E1007 15:01:00.152249 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63f04186-4fb9-4518-b19e-b0a7bf573df0" containerName="collect-profiles" Oct 07 15:01:00 crc kubenswrapper[4959]: I1007 15:01:00.152267 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="63f04186-4fb9-4518-b19e-b0a7bf573df0" containerName="collect-profiles" Oct 07 15:01:00 crc kubenswrapper[4959]: I1007 15:01:00.152514 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="63f04186-4fb9-4518-b19e-b0a7bf573df0" containerName="collect-profiles" Oct 07 15:01:00 crc kubenswrapper[4959]: I1007 15:01:00.153411 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29330821-dpxgj" Oct 07 15:01:00 crc kubenswrapper[4959]: I1007 15:01:00.185239 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29330821-dpxgj"] Oct 07 15:01:00 crc kubenswrapper[4959]: I1007 15:01:00.282258 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e167acc8-aaaa-4595-8543-a9622ee92fc8-combined-ca-bundle\") pod \"keystone-cron-29330821-dpxgj\" (UID: \"e167acc8-aaaa-4595-8543-a9622ee92fc8\") " pod="openstack/keystone-cron-29330821-dpxgj" Oct 07 15:01:00 crc kubenswrapper[4959]: I1007 15:01:00.282324 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e167acc8-aaaa-4595-8543-a9622ee92fc8-fernet-keys\") pod \"keystone-cron-29330821-dpxgj\" (UID: \"e167acc8-aaaa-4595-8543-a9622ee92fc8\") " pod="openstack/keystone-cron-29330821-dpxgj" Oct 07 15:01:00 crc kubenswrapper[4959]: I1007 15:01:00.282498 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e167acc8-aaaa-4595-8543-a9622ee92fc8-config-data\") pod \"keystone-cron-29330821-dpxgj\" (UID: \"e167acc8-aaaa-4595-8543-a9622ee92fc8\") " pod="openstack/keystone-cron-29330821-dpxgj" Oct 07 15:01:00 crc kubenswrapper[4959]: I1007 15:01:00.282529 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jqxm\" (UniqueName: \"kubernetes.io/projected/e167acc8-aaaa-4595-8543-a9622ee92fc8-kube-api-access-5jqxm\") pod \"keystone-cron-29330821-dpxgj\" (UID: \"e167acc8-aaaa-4595-8543-a9622ee92fc8\") " pod="openstack/keystone-cron-29330821-dpxgj" Oct 07 15:01:00 crc kubenswrapper[4959]: I1007 15:01:00.384900 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e167acc8-aaaa-4595-8543-a9622ee92fc8-config-data\") pod \"keystone-cron-29330821-dpxgj\" (UID: \"e167acc8-aaaa-4595-8543-a9622ee92fc8\") " pod="openstack/keystone-cron-29330821-dpxgj" Oct 07 15:01:00 crc kubenswrapper[4959]: I1007 15:01:00.384968 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jqxm\" (UniqueName: \"kubernetes.io/projected/e167acc8-aaaa-4595-8543-a9622ee92fc8-kube-api-access-5jqxm\") pod \"keystone-cron-29330821-dpxgj\" (UID: \"e167acc8-aaaa-4595-8543-a9622ee92fc8\") " pod="openstack/keystone-cron-29330821-dpxgj" Oct 07 15:01:00 crc kubenswrapper[4959]: I1007 15:01:00.385121 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e167acc8-aaaa-4595-8543-a9622ee92fc8-combined-ca-bundle\") pod \"keystone-cron-29330821-dpxgj\" (UID: \"e167acc8-aaaa-4595-8543-a9622ee92fc8\") " pod="openstack/keystone-cron-29330821-dpxgj" Oct 07 15:01:00 crc kubenswrapper[4959]: I1007 15:01:00.385149 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e167acc8-aaaa-4595-8543-a9622ee92fc8-fernet-keys\") pod \"keystone-cron-29330821-dpxgj\" (UID: \"e167acc8-aaaa-4595-8543-a9622ee92fc8\") " pod="openstack/keystone-cron-29330821-dpxgj" Oct 07 15:01:00 crc kubenswrapper[4959]: I1007 15:01:00.395234 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e167acc8-aaaa-4595-8543-a9622ee92fc8-config-data\") pod \"keystone-cron-29330821-dpxgj\" (UID: \"e167acc8-aaaa-4595-8543-a9622ee92fc8\") " pod="openstack/keystone-cron-29330821-dpxgj" Oct 07 15:01:00 crc kubenswrapper[4959]: I1007 15:01:00.399016 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e167acc8-aaaa-4595-8543-a9622ee92fc8-combined-ca-bundle\") pod \"keystone-cron-29330821-dpxgj\" (UID: \"e167acc8-aaaa-4595-8543-a9622ee92fc8\") " pod="openstack/keystone-cron-29330821-dpxgj" Oct 07 15:01:00 crc kubenswrapper[4959]: I1007 15:01:00.400699 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e167acc8-aaaa-4595-8543-a9622ee92fc8-fernet-keys\") pod \"keystone-cron-29330821-dpxgj\" (UID: \"e167acc8-aaaa-4595-8543-a9622ee92fc8\") " pod="openstack/keystone-cron-29330821-dpxgj" Oct 07 15:01:00 crc kubenswrapper[4959]: I1007 15:01:00.406532 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jqxm\" (UniqueName: \"kubernetes.io/projected/e167acc8-aaaa-4595-8543-a9622ee92fc8-kube-api-access-5jqxm\") pod \"keystone-cron-29330821-dpxgj\" (UID: \"e167acc8-aaaa-4595-8543-a9622ee92fc8\") " pod="openstack/keystone-cron-29330821-dpxgj" Oct 07 15:01:00 crc kubenswrapper[4959]: I1007 15:01:00.479454 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29330821-dpxgj" Oct 07 15:01:01 crc kubenswrapper[4959]: I1007 15:01:01.031386 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29330821-dpxgj"] Oct 07 15:01:01 crc kubenswrapper[4959]: I1007 15:01:01.298042 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29330821-dpxgj" event={"ID":"e167acc8-aaaa-4595-8543-a9622ee92fc8","Type":"ContainerStarted","Data":"66eb409b2121fe3cbe45de1307feba745b8cd4a8fcc8af4ecacdf3c3ce52508f"} Oct 07 15:01:02 crc kubenswrapper[4959]: I1007 15:01:02.310094 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29330821-dpxgj" event={"ID":"e167acc8-aaaa-4595-8543-a9622ee92fc8","Type":"ContainerStarted","Data":"2732341dd74670b06e2e419fda279d950bfbbdf069dd578a8edb75b231af2aee"} Oct 07 15:01:02 crc kubenswrapper[4959]: I1007 15:01:02.329140 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29330821-dpxgj" podStartSLOduration=2.329117975 podStartE2EDuration="2.329117975s" podCreationTimestamp="2025-10-07 15:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 15:01:02.325956655 +0000 UTC m=+4544.409360980" watchObservedRunningTime="2025-10-07 15:01:02.329117975 +0000 UTC m=+4544.412522300" Oct 07 15:01:05 crc kubenswrapper[4959]: I1007 15:01:05.345618 4959 generic.go:334] "Generic (PLEG): container finished" podID="e167acc8-aaaa-4595-8543-a9622ee92fc8" containerID="2732341dd74670b06e2e419fda279d950bfbbdf069dd578a8edb75b231af2aee" exitCode=0 Oct 07 15:01:05 crc kubenswrapper[4959]: I1007 15:01:05.346418 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29330821-dpxgj" event={"ID":"e167acc8-aaaa-4595-8543-a9622ee92fc8","Type":"ContainerDied","Data":"2732341dd74670b06e2e419fda279d950bfbbdf069dd578a8edb75b231af2aee"} Oct 07 15:01:06 crc kubenswrapper[4959]: I1007 15:01:06.870425 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29330821-dpxgj" Oct 07 15:01:06 crc kubenswrapper[4959]: I1007 15:01:06.949510 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e167acc8-aaaa-4595-8543-a9622ee92fc8-combined-ca-bundle\") pod \"e167acc8-aaaa-4595-8543-a9622ee92fc8\" (UID: \"e167acc8-aaaa-4595-8543-a9622ee92fc8\") " Oct 07 15:01:06 crc kubenswrapper[4959]: I1007 15:01:06.949611 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e167acc8-aaaa-4595-8543-a9622ee92fc8-config-data\") pod \"e167acc8-aaaa-4595-8543-a9622ee92fc8\" (UID: \"e167acc8-aaaa-4595-8543-a9622ee92fc8\") " Oct 07 15:01:06 crc kubenswrapper[4959]: I1007 15:01:06.949781 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e167acc8-aaaa-4595-8543-a9622ee92fc8-fernet-keys\") pod \"e167acc8-aaaa-4595-8543-a9622ee92fc8\" (UID: \"e167acc8-aaaa-4595-8543-a9622ee92fc8\") " Oct 07 15:01:06 crc kubenswrapper[4959]: I1007 15:01:06.949874 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jqxm\" (UniqueName: \"kubernetes.io/projected/e167acc8-aaaa-4595-8543-a9622ee92fc8-kube-api-access-5jqxm\") pod \"e167acc8-aaaa-4595-8543-a9622ee92fc8\" (UID: \"e167acc8-aaaa-4595-8543-a9622ee92fc8\") " Oct 07 15:01:06 crc kubenswrapper[4959]: I1007 15:01:06.958598 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e167acc8-aaaa-4595-8543-a9622ee92fc8-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "e167acc8-aaaa-4595-8543-a9622ee92fc8" (UID: "e167acc8-aaaa-4595-8543-a9622ee92fc8"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:01:06 crc kubenswrapper[4959]: I1007 15:01:06.960585 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e167acc8-aaaa-4595-8543-a9622ee92fc8-kube-api-access-5jqxm" (OuterVolumeSpecName: "kube-api-access-5jqxm") pod "e167acc8-aaaa-4595-8543-a9622ee92fc8" (UID: "e167acc8-aaaa-4595-8543-a9622ee92fc8"). InnerVolumeSpecName "kube-api-access-5jqxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 15:01:06 crc kubenswrapper[4959]: I1007 15:01:06.987434 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e167acc8-aaaa-4595-8543-a9622ee92fc8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e167acc8-aaaa-4595-8543-a9622ee92fc8" (UID: "e167acc8-aaaa-4595-8543-a9622ee92fc8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:01:07 crc kubenswrapper[4959]: I1007 15:01:07.033240 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e167acc8-aaaa-4595-8543-a9622ee92fc8-config-data" (OuterVolumeSpecName: "config-data") pod "e167acc8-aaaa-4595-8543-a9622ee92fc8" (UID: "e167acc8-aaaa-4595-8543-a9622ee92fc8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:01:07 crc kubenswrapper[4959]: I1007 15:01:07.053404 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e167acc8-aaaa-4595-8543-a9622ee92fc8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 07 15:01:07 crc kubenswrapper[4959]: I1007 15:01:07.053728 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e167acc8-aaaa-4595-8543-a9622ee92fc8-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 15:01:07 crc kubenswrapper[4959]: I1007 15:01:07.053817 4959 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e167acc8-aaaa-4595-8543-a9622ee92fc8-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 07 15:01:07 crc kubenswrapper[4959]: I1007 15:01:07.053959 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jqxm\" (UniqueName: \"kubernetes.io/projected/e167acc8-aaaa-4595-8543-a9622ee92fc8-kube-api-access-5jqxm\") on node \"crc\" DevicePath \"\"" Oct 07 15:01:07 crc kubenswrapper[4959]: I1007 15:01:07.369816 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29330821-dpxgj" event={"ID":"e167acc8-aaaa-4595-8543-a9622ee92fc8","Type":"ContainerDied","Data":"66eb409b2121fe3cbe45de1307feba745b8cd4a8fcc8af4ecacdf3c3ce52508f"} Oct 07 15:01:07 crc kubenswrapper[4959]: I1007 15:01:07.369882 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66eb409b2121fe3cbe45de1307feba745b8cd4a8fcc8af4ecacdf3c3ce52508f" Oct 07 15:01:07 crc kubenswrapper[4959]: I1007 15:01:07.369924 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29330821-dpxgj" Oct 07 15:01:08 crc kubenswrapper[4959]: I1007 15:01:08.661399 4959 scope.go:117] "RemoveContainer" containerID="9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204" Oct 07 15:01:08 crc kubenswrapper[4959]: E1007 15:01:08.662251 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:01:20 crc kubenswrapper[4959]: I1007 15:01:20.653871 4959 scope.go:117] "RemoveContainer" containerID="9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204" Oct 07 15:01:20 crc kubenswrapper[4959]: E1007 15:01:20.654629 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:01:32 crc kubenswrapper[4959]: I1007 15:01:32.654907 4959 scope.go:117] "RemoveContainer" containerID="9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204" Oct 07 15:01:32 crc kubenswrapper[4959]: E1007 15:01:32.656410 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:01:43 crc kubenswrapper[4959]: I1007 15:01:43.654506 4959 scope.go:117] "RemoveContainer" containerID="9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204" Oct 07 15:01:43 crc kubenswrapper[4959]: E1007 15:01:43.655818 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:01:58 crc kubenswrapper[4959]: I1007 15:01:58.661522 4959 scope.go:117] "RemoveContainer" containerID="9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204" Oct 07 15:01:58 crc kubenswrapper[4959]: E1007 15:01:58.662941 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:02:13 crc kubenswrapper[4959]: I1007 15:02:13.654312 4959 scope.go:117] "RemoveContainer" containerID="9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204" Oct 07 15:02:13 crc kubenswrapper[4959]: E1007 15:02:13.655239 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:02:25 crc kubenswrapper[4959]: I1007 15:02:25.653630 4959 scope.go:117] "RemoveContainer" containerID="9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204" Oct 07 15:02:26 crc kubenswrapper[4959]: I1007 15:02:26.123799 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerStarted","Data":"e136903417b368464eb103f0eb291b140d45a4b3861749d2f8a77717fa3714ce"} Oct 07 15:04:23 crc kubenswrapper[4959]: I1007 15:04:23.200118 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dfp5q"] Oct 07 15:04:23 crc kubenswrapper[4959]: E1007 15:04:23.201280 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e167acc8-aaaa-4595-8543-a9622ee92fc8" containerName="keystone-cron" Oct 07 15:04:23 crc kubenswrapper[4959]: I1007 15:04:23.201298 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e167acc8-aaaa-4595-8543-a9622ee92fc8" containerName="keystone-cron" Oct 07 15:04:23 crc kubenswrapper[4959]: I1007 15:04:23.201539 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="e167acc8-aaaa-4595-8543-a9622ee92fc8" containerName="keystone-cron" Oct 07 15:04:23 crc kubenswrapper[4959]: I1007 15:04:23.203355 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dfp5q" Oct 07 15:04:23 crc kubenswrapper[4959]: I1007 15:04:23.230274 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dfp5q"] Oct 07 15:04:23 crc kubenswrapper[4959]: I1007 15:04:23.290046 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzwbp\" (UniqueName: \"kubernetes.io/projected/da33c6e1-4e26-4b27-908c-962a398740f0-kube-api-access-rzwbp\") pod \"redhat-operators-dfp5q\" (UID: \"da33c6e1-4e26-4b27-908c-962a398740f0\") " pod="openshift-marketplace/redhat-operators-dfp5q" Oct 07 15:04:23 crc kubenswrapper[4959]: I1007 15:04:23.290176 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da33c6e1-4e26-4b27-908c-962a398740f0-utilities\") pod \"redhat-operators-dfp5q\" (UID: \"da33c6e1-4e26-4b27-908c-962a398740f0\") " pod="openshift-marketplace/redhat-operators-dfp5q" Oct 07 15:04:23 crc kubenswrapper[4959]: I1007 15:04:23.290279 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da33c6e1-4e26-4b27-908c-962a398740f0-catalog-content\") pod \"redhat-operators-dfp5q\" (UID: \"da33c6e1-4e26-4b27-908c-962a398740f0\") " pod="openshift-marketplace/redhat-operators-dfp5q" Oct 07 15:04:23 crc kubenswrapper[4959]: I1007 15:04:23.396295 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da33c6e1-4e26-4b27-908c-962a398740f0-utilities\") pod \"redhat-operators-dfp5q\" (UID: \"da33c6e1-4e26-4b27-908c-962a398740f0\") " pod="openshift-marketplace/redhat-operators-dfp5q" Oct 07 15:04:23 crc kubenswrapper[4959]: I1007 15:04:23.396485 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da33c6e1-4e26-4b27-908c-962a398740f0-catalog-content\") pod \"redhat-operators-dfp5q\" (UID: \"da33c6e1-4e26-4b27-908c-962a398740f0\") " pod="openshift-marketplace/redhat-operators-dfp5q" Oct 07 15:04:23 crc kubenswrapper[4959]: I1007 15:04:23.396729 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzwbp\" (UniqueName: \"kubernetes.io/projected/da33c6e1-4e26-4b27-908c-962a398740f0-kube-api-access-rzwbp\") pod \"redhat-operators-dfp5q\" (UID: \"da33c6e1-4e26-4b27-908c-962a398740f0\") " pod="openshift-marketplace/redhat-operators-dfp5q" Oct 07 15:04:23 crc kubenswrapper[4959]: I1007 15:04:23.397713 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da33c6e1-4e26-4b27-908c-962a398740f0-utilities\") pod \"redhat-operators-dfp5q\" (UID: \"da33c6e1-4e26-4b27-908c-962a398740f0\") " pod="openshift-marketplace/redhat-operators-dfp5q" Oct 07 15:04:23 crc kubenswrapper[4959]: I1007 15:04:23.398022 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da33c6e1-4e26-4b27-908c-962a398740f0-catalog-content\") pod \"redhat-operators-dfp5q\" (UID: \"da33c6e1-4e26-4b27-908c-962a398740f0\") " pod="openshift-marketplace/redhat-operators-dfp5q" Oct 07 15:04:23 crc kubenswrapper[4959]: I1007 15:04:23.415081 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzwbp\" (UniqueName: \"kubernetes.io/projected/da33c6e1-4e26-4b27-908c-962a398740f0-kube-api-access-rzwbp\") pod \"redhat-operators-dfp5q\" (UID: \"da33c6e1-4e26-4b27-908c-962a398740f0\") " pod="openshift-marketplace/redhat-operators-dfp5q" Oct 07 15:04:23 crc kubenswrapper[4959]: I1007 15:04:23.551384 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dfp5q" Oct 07 15:04:24 crc kubenswrapper[4959]: I1007 15:04:24.120178 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dfp5q"] Oct 07 15:04:24 crc kubenswrapper[4959]: I1007 15:04:24.304320 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfp5q" event={"ID":"da33c6e1-4e26-4b27-908c-962a398740f0","Type":"ContainerStarted","Data":"317c2abcb949f4561a3fa8688785ee78ed5101e0797d07e2801b6b2498afd040"} Oct 07 15:04:25 crc kubenswrapper[4959]: I1007 15:04:25.315654 4959 generic.go:334] "Generic (PLEG): container finished" podID="da33c6e1-4e26-4b27-908c-962a398740f0" containerID="5a4a50b7169377ff70c20f42fc9f755090769712ef64c162b334861c86138bd5" exitCode=0 Oct 07 15:04:25 crc kubenswrapper[4959]: I1007 15:04:25.315913 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfp5q" event={"ID":"da33c6e1-4e26-4b27-908c-962a398740f0","Type":"ContainerDied","Data":"5a4a50b7169377ff70c20f42fc9f755090769712ef64c162b334861c86138bd5"} Oct 07 15:04:25 crc kubenswrapper[4959]: I1007 15:04:25.318678 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 15:04:27 crc kubenswrapper[4959]: I1007 15:04:27.337599 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfp5q" event={"ID":"da33c6e1-4e26-4b27-908c-962a398740f0","Type":"ContainerStarted","Data":"45a30c2f536e1a2a8c92a2996f9ed046a122665cd2d8d57c6eb3e2014c24ae37"} Oct 07 15:04:34 crc kubenswrapper[4959]: I1007 15:04:34.397456 4959 generic.go:334] "Generic (PLEG): container finished" podID="da33c6e1-4e26-4b27-908c-962a398740f0" containerID="45a30c2f536e1a2a8c92a2996f9ed046a122665cd2d8d57c6eb3e2014c24ae37" exitCode=0 Oct 07 15:04:34 crc kubenswrapper[4959]: I1007 15:04:34.397534 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfp5q" event={"ID":"da33c6e1-4e26-4b27-908c-962a398740f0","Type":"ContainerDied","Data":"45a30c2f536e1a2a8c92a2996f9ed046a122665cd2d8d57c6eb3e2014c24ae37"} Oct 07 15:04:35 crc kubenswrapper[4959]: I1007 15:04:35.411338 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfp5q" event={"ID":"da33c6e1-4e26-4b27-908c-962a398740f0","Type":"ContainerStarted","Data":"aaeaecefdf7822cf4da8e79faa890e54fd6f90e9a1415a4d1b21319fd62e617d"} Oct 07 15:04:35 crc kubenswrapper[4959]: I1007 15:04:35.447748 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dfp5q" podStartSLOduration=2.787211886 podStartE2EDuration="12.447720179s" podCreationTimestamp="2025-10-07 15:04:23 +0000 UTC" firstStartedPulling="2025-10-07 15:04:25.318473386 +0000 UTC m=+4747.401877711" lastFinishedPulling="2025-10-07 15:04:34.978981679 +0000 UTC m=+4757.062386004" observedRunningTime="2025-10-07 15:04:35.438163105 +0000 UTC m=+4757.521567430" watchObservedRunningTime="2025-10-07 15:04:35.447720179 +0000 UTC m=+4757.531124504" Oct 07 15:04:43 crc kubenswrapper[4959]: I1007 15:04:43.552703 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dfp5q" Oct 07 15:04:43 crc kubenswrapper[4959]: I1007 15:04:43.553881 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dfp5q" Oct 07 15:04:44 crc kubenswrapper[4959]: I1007 15:04:44.596526 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dfp5q" podUID="da33c6e1-4e26-4b27-908c-962a398740f0" containerName="registry-server" probeResult="failure" output=< Oct 07 15:04:44 crc kubenswrapper[4959]: timeout: failed to connect service ":50051" within 1s Oct 07 15:04:44 crc kubenswrapper[4959]: > Oct 07 15:04:50 crc kubenswrapper[4959]: I1007 15:04:50.630820 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 15:04:50 crc kubenswrapper[4959]: I1007 15:04:50.631829 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 15:04:53 crc kubenswrapper[4959]: I1007 15:04:53.603597 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dfp5q" Oct 07 15:04:53 crc kubenswrapper[4959]: I1007 15:04:53.657826 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dfp5q" Oct 07 15:04:54 crc kubenswrapper[4959]: I1007 15:04:54.408214 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dfp5q"] Oct 07 15:04:55 crc kubenswrapper[4959]: I1007 15:04:55.586945 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dfp5q" podUID="da33c6e1-4e26-4b27-908c-962a398740f0" containerName="registry-server" containerID="cri-o://aaeaecefdf7822cf4da8e79faa890e54fd6f90e9a1415a4d1b21319fd62e617d" gracePeriod=2 Oct 07 15:04:56 crc kubenswrapper[4959]: I1007 15:04:56.212045 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dfp5q" Oct 07 15:04:56 crc kubenswrapper[4959]: I1007 15:04:56.359674 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da33c6e1-4e26-4b27-908c-962a398740f0-catalog-content\") pod \"da33c6e1-4e26-4b27-908c-962a398740f0\" (UID: \"da33c6e1-4e26-4b27-908c-962a398740f0\") " Oct 07 15:04:56 crc kubenswrapper[4959]: I1007 15:04:56.360067 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzwbp\" (UniqueName: \"kubernetes.io/projected/da33c6e1-4e26-4b27-908c-962a398740f0-kube-api-access-rzwbp\") pod \"da33c6e1-4e26-4b27-908c-962a398740f0\" (UID: \"da33c6e1-4e26-4b27-908c-962a398740f0\") " Oct 07 15:04:56 crc kubenswrapper[4959]: I1007 15:04:56.360166 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da33c6e1-4e26-4b27-908c-962a398740f0-utilities\") pod \"da33c6e1-4e26-4b27-908c-962a398740f0\" (UID: \"da33c6e1-4e26-4b27-908c-962a398740f0\") " Oct 07 15:04:56 crc kubenswrapper[4959]: I1007 15:04:56.361333 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da33c6e1-4e26-4b27-908c-962a398740f0-utilities" (OuterVolumeSpecName: "utilities") pod "da33c6e1-4e26-4b27-908c-962a398740f0" (UID: "da33c6e1-4e26-4b27-908c-962a398740f0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:04:56 crc kubenswrapper[4959]: I1007 15:04:56.365461 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da33c6e1-4e26-4b27-908c-962a398740f0-kube-api-access-rzwbp" (OuterVolumeSpecName: "kube-api-access-rzwbp") pod "da33c6e1-4e26-4b27-908c-962a398740f0" (UID: "da33c6e1-4e26-4b27-908c-962a398740f0"). InnerVolumeSpecName "kube-api-access-rzwbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 15:04:56 crc kubenswrapper[4959]: I1007 15:04:56.454961 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da33c6e1-4e26-4b27-908c-962a398740f0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "da33c6e1-4e26-4b27-908c-962a398740f0" (UID: "da33c6e1-4e26-4b27-908c-962a398740f0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:04:56 crc kubenswrapper[4959]: I1007 15:04:56.462744 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da33c6e1-4e26-4b27-908c-962a398740f0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 15:04:56 crc kubenswrapper[4959]: I1007 15:04:56.462782 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzwbp\" (UniqueName: \"kubernetes.io/projected/da33c6e1-4e26-4b27-908c-962a398740f0-kube-api-access-rzwbp\") on node \"crc\" DevicePath \"\"" Oct 07 15:04:56 crc kubenswrapper[4959]: I1007 15:04:56.462796 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da33c6e1-4e26-4b27-908c-962a398740f0-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 15:04:56 crc kubenswrapper[4959]: I1007 15:04:56.597483 4959 generic.go:334] "Generic (PLEG): container finished" podID="da33c6e1-4e26-4b27-908c-962a398740f0" containerID="aaeaecefdf7822cf4da8e79faa890e54fd6f90e9a1415a4d1b21319fd62e617d" exitCode=0 Oct 07 15:04:56 crc kubenswrapper[4959]: I1007 15:04:56.597530 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfp5q" event={"ID":"da33c6e1-4e26-4b27-908c-962a398740f0","Type":"ContainerDied","Data":"aaeaecefdf7822cf4da8e79faa890e54fd6f90e9a1415a4d1b21319fd62e617d"} Oct 07 15:04:56 crc kubenswrapper[4959]: I1007 15:04:56.597559 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfp5q" event={"ID":"da33c6e1-4e26-4b27-908c-962a398740f0","Type":"ContainerDied","Data":"317c2abcb949f4561a3fa8688785ee78ed5101e0797d07e2801b6b2498afd040"} Oct 07 15:04:56 crc kubenswrapper[4959]: I1007 15:04:56.597576 4959 scope.go:117] "RemoveContainer" containerID="aaeaecefdf7822cf4da8e79faa890e54fd6f90e9a1415a4d1b21319fd62e617d" Oct 07 15:04:56 crc kubenswrapper[4959]: I1007 15:04:56.597710 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dfp5q" Oct 07 15:04:56 crc kubenswrapper[4959]: I1007 15:04:56.634127 4959 scope.go:117] "RemoveContainer" containerID="45a30c2f536e1a2a8c92a2996f9ed046a122665cd2d8d57c6eb3e2014c24ae37" Oct 07 15:04:56 crc kubenswrapper[4959]: I1007 15:04:56.635586 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dfp5q"] Oct 07 15:04:56 crc kubenswrapper[4959]: I1007 15:04:56.647081 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dfp5q"] Oct 07 15:04:56 crc kubenswrapper[4959]: I1007 15:04:56.669149 4959 scope.go:117] "RemoveContainer" containerID="5a4a50b7169377ff70c20f42fc9f755090769712ef64c162b334861c86138bd5" Oct 07 15:04:56 crc kubenswrapper[4959]: I1007 15:04:56.696738 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da33c6e1-4e26-4b27-908c-962a398740f0" path="/var/lib/kubelet/pods/da33c6e1-4e26-4b27-908c-962a398740f0/volumes" Oct 07 15:04:56 crc kubenswrapper[4959]: I1007 15:04:56.711734 4959 scope.go:117] "RemoveContainer" containerID="aaeaecefdf7822cf4da8e79faa890e54fd6f90e9a1415a4d1b21319fd62e617d" Oct 07 15:04:56 crc kubenswrapper[4959]: E1007 15:04:56.712466 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aaeaecefdf7822cf4da8e79faa890e54fd6f90e9a1415a4d1b21319fd62e617d\": container with ID starting with aaeaecefdf7822cf4da8e79faa890e54fd6f90e9a1415a4d1b21319fd62e617d not found: ID does not exist" containerID="aaeaecefdf7822cf4da8e79faa890e54fd6f90e9a1415a4d1b21319fd62e617d" Oct 07 15:04:56 crc kubenswrapper[4959]: I1007 15:04:56.712557 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aaeaecefdf7822cf4da8e79faa890e54fd6f90e9a1415a4d1b21319fd62e617d"} err="failed to get container status \"aaeaecefdf7822cf4da8e79faa890e54fd6f90e9a1415a4d1b21319fd62e617d\": rpc error: code = NotFound desc = could not find container \"aaeaecefdf7822cf4da8e79faa890e54fd6f90e9a1415a4d1b21319fd62e617d\": container with ID starting with aaeaecefdf7822cf4da8e79faa890e54fd6f90e9a1415a4d1b21319fd62e617d not found: ID does not exist" Oct 07 15:04:56 crc kubenswrapper[4959]: I1007 15:04:56.712591 4959 scope.go:117] "RemoveContainer" containerID="45a30c2f536e1a2a8c92a2996f9ed046a122665cd2d8d57c6eb3e2014c24ae37" Oct 07 15:04:56 crc kubenswrapper[4959]: E1007 15:04:56.713202 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45a30c2f536e1a2a8c92a2996f9ed046a122665cd2d8d57c6eb3e2014c24ae37\": container with ID starting with 45a30c2f536e1a2a8c92a2996f9ed046a122665cd2d8d57c6eb3e2014c24ae37 not found: ID does not exist" containerID="45a30c2f536e1a2a8c92a2996f9ed046a122665cd2d8d57c6eb3e2014c24ae37" Oct 07 15:04:56 crc kubenswrapper[4959]: I1007 15:04:56.713249 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45a30c2f536e1a2a8c92a2996f9ed046a122665cd2d8d57c6eb3e2014c24ae37"} err="failed to get container status \"45a30c2f536e1a2a8c92a2996f9ed046a122665cd2d8d57c6eb3e2014c24ae37\": rpc error: code = NotFound desc = could not find container \"45a30c2f536e1a2a8c92a2996f9ed046a122665cd2d8d57c6eb3e2014c24ae37\": container with ID starting with 45a30c2f536e1a2a8c92a2996f9ed046a122665cd2d8d57c6eb3e2014c24ae37 not found: ID does not exist" Oct 07 15:04:56 crc kubenswrapper[4959]: I1007 15:04:56.713284 4959 scope.go:117] "RemoveContainer" containerID="5a4a50b7169377ff70c20f42fc9f755090769712ef64c162b334861c86138bd5" Oct 07 15:04:56 crc kubenswrapper[4959]: E1007 15:04:56.714132 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a4a50b7169377ff70c20f42fc9f755090769712ef64c162b334861c86138bd5\": container with ID starting with 5a4a50b7169377ff70c20f42fc9f755090769712ef64c162b334861c86138bd5 not found: ID does not exist" containerID="5a4a50b7169377ff70c20f42fc9f755090769712ef64c162b334861c86138bd5" Oct 07 15:04:56 crc kubenswrapper[4959]: I1007 15:04:56.714166 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a4a50b7169377ff70c20f42fc9f755090769712ef64c162b334861c86138bd5"} err="failed to get container status \"5a4a50b7169377ff70c20f42fc9f755090769712ef64c162b334861c86138bd5\": rpc error: code = NotFound desc = could not find container \"5a4a50b7169377ff70c20f42fc9f755090769712ef64c162b334861c86138bd5\": container with ID starting with 5a4a50b7169377ff70c20f42fc9f755090769712ef64c162b334861c86138bd5 not found: ID does not exist" Oct 07 15:05:20 crc kubenswrapper[4959]: I1007 15:05:20.630070 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 15:05:20 crc kubenswrapper[4959]: I1007 15:05:20.631139 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 15:05:39 crc kubenswrapper[4959]: I1007 15:05:39.374045 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rphjb"] Oct 07 15:05:39 crc kubenswrapper[4959]: E1007 15:05:39.375260 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da33c6e1-4e26-4b27-908c-962a398740f0" containerName="registry-server" Oct 07 15:05:39 crc kubenswrapper[4959]: I1007 15:05:39.375278 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="da33c6e1-4e26-4b27-908c-962a398740f0" containerName="registry-server" Oct 07 15:05:39 crc kubenswrapper[4959]: E1007 15:05:39.375300 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da33c6e1-4e26-4b27-908c-962a398740f0" containerName="extract-content" Oct 07 15:05:39 crc kubenswrapper[4959]: I1007 15:05:39.375309 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="da33c6e1-4e26-4b27-908c-962a398740f0" containerName="extract-content" Oct 07 15:05:39 crc kubenswrapper[4959]: E1007 15:05:39.375323 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da33c6e1-4e26-4b27-908c-962a398740f0" containerName="extract-utilities" Oct 07 15:05:39 crc kubenswrapper[4959]: I1007 15:05:39.375331 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="da33c6e1-4e26-4b27-908c-962a398740f0" containerName="extract-utilities" Oct 07 15:05:39 crc kubenswrapper[4959]: I1007 15:05:39.375615 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="da33c6e1-4e26-4b27-908c-962a398740f0" containerName="registry-server" Oct 07 15:05:39 crc kubenswrapper[4959]: I1007 15:05:39.377500 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rphjb" Oct 07 15:05:39 crc kubenswrapper[4959]: I1007 15:05:39.387315 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rphjb"] Oct 07 15:05:39 crc kubenswrapper[4959]: I1007 15:05:39.462228 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvrf5\" (UniqueName: \"kubernetes.io/projected/2ea7afff-d8c0-4ecd-9a84-17049000c904-kube-api-access-cvrf5\") pod \"redhat-marketplace-rphjb\" (UID: \"2ea7afff-d8c0-4ecd-9a84-17049000c904\") " pod="openshift-marketplace/redhat-marketplace-rphjb" Oct 07 15:05:39 crc kubenswrapper[4959]: I1007 15:05:39.462312 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ea7afff-d8c0-4ecd-9a84-17049000c904-catalog-content\") pod \"redhat-marketplace-rphjb\" (UID: \"2ea7afff-d8c0-4ecd-9a84-17049000c904\") " pod="openshift-marketplace/redhat-marketplace-rphjb" Oct 07 15:05:39 crc kubenswrapper[4959]: I1007 15:05:39.462407 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ea7afff-d8c0-4ecd-9a84-17049000c904-utilities\") pod \"redhat-marketplace-rphjb\" (UID: \"2ea7afff-d8c0-4ecd-9a84-17049000c904\") " pod="openshift-marketplace/redhat-marketplace-rphjb" Oct 07 15:05:39 crc kubenswrapper[4959]: I1007 15:05:39.565373 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvrf5\" (UniqueName: \"kubernetes.io/projected/2ea7afff-d8c0-4ecd-9a84-17049000c904-kube-api-access-cvrf5\") pod \"redhat-marketplace-rphjb\" (UID: \"2ea7afff-d8c0-4ecd-9a84-17049000c904\") " pod="openshift-marketplace/redhat-marketplace-rphjb" Oct 07 15:05:39 crc kubenswrapper[4959]: I1007 15:05:39.565447 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ea7afff-d8c0-4ecd-9a84-17049000c904-catalog-content\") pod \"redhat-marketplace-rphjb\" (UID: \"2ea7afff-d8c0-4ecd-9a84-17049000c904\") " pod="openshift-marketplace/redhat-marketplace-rphjb" Oct 07 15:05:39 crc kubenswrapper[4959]: I1007 15:05:39.565520 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ea7afff-d8c0-4ecd-9a84-17049000c904-utilities\") pod \"redhat-marketplace-rphjb\" (UID: \"2ea7afff-d8c0-4ecd-9a84-17049000c904\") " pod="openshift-marketplace/redhat-marketplace-rphjb" Oct 07 15:05:39 crc kubenswrapper[4959]: I1007 15:05:39.566248 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ea7afff-d8c0-4ecd-9a84-17049000c904-utilities\") pod \"redhat-marketplace-rphjb\" (UID: \"2ea7afff-d8c0-4ecd-9a84-17049000c904\") " pod="openshift-marketplace/redhat-marketplace-rphjb" Oct 07 15:05:39 crc kubenswrapper[4959]: I1007 15:05:39.566290 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ea7afff-d8c0-4ecd-9a84-17049000c904-catalog-content\") pod \"redhat-marketplace-rphjb\" (UID: \"2ea7afff-d8c0-4ecd-9a84-17049000c904\") " pod="openshift-marketplace/redhat-marketplace-rphjb" Oct 07 15:05:39 crc kubenswrapper[4959]: I1007 15:05:39.596405 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvrf5\" (UniqueName: \"kubernetes.io/projected/2ea7afff-d8c0-4ecd-9a84-17049000c904-kube-api-access-cvrf5\") pod \"redhat-marketplace-rphjb\" (UID: \"2ea7afff-d8c0-4ecd-9a84-17049000c904\") " pod="openshift-marketplace/redhat-marketplace-rphjb" Oct 07 15:05:39 crc kubenswrapper[4959]: I1007 15:05:39.702551 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rphjb" Oct 07 15:05:40 crc kubenswrapper[4959]: I1007 15:05:40.207367 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rphjb"] Oct 07 15:05:41 crc kubenswrapper[4959]: I1007 15:05:41.076026 4959 generic.go:334] "Generic (PLEG): container finished" podID="2ea7afff-d8c0-4ecd-9a84-17049000c904" containerID="5fcc03a706183c519c10f3d7d49240d31b3c0319189ab9b57a20421158bfdd48" exitCode=0 Oct 07 15:05:41 crc kubenswrapper[4959]: I1007 15:05:41.076506 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rphjb" event={"ID":"2ea7afff-d8c0-4ecd-9a84-17049000c904","Type":"ContainerDied","Data":"5fcc03a706183c519c10f3d7d49240d31b3c0319189ab9b57a20421158bfdd48"} Oct 07 15:05:41 crc kubenswrapper[4959]: I1007 15:05:41.076551 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rphjb" event={"ID":"2ea7afff-d8c0-4ecd-9a84-17049000c904","Type":"ContainerStarted","Data":"ff4867a9ea8753a8e62d7b8c5f1d9ff7f4bb09a9ab5a4dad124d5509ed32570e"} Oct 07 15:05:43 crc kubenswrapper[4959]: I1007 15:05:43.099629 4959 generic.go:334] "Generic (PLEG): container finished" podID="2ea7afff-d8c0-4ecd-9a84-17049000c904" containerID="c3003a57ba13d8aa519ecc51f8c5260b12ef8c8a88876a2f14ceccc97bed6ca7" exitCode=0 Oct 07 15:05:43 crc kubenswrapper[4959]: I1007 15:05:43.099771 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rphjb" event={"ID":"2ea7afff-d8c0-4ecd-9a84-17049000c904","Type":"ContainerDied","Data":"c3003a57ba13d8aa519ecc51f8c5260b12ef8c8a88876a2f14ceccc97bed6ca7"} Oct 07 15:05:44 crc kubenswrapper[4959]: I1007 15:05:44.116216 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rphjb" event={"ID":"2ea7afff-d8c0-4ecd-9a84-17049000c904","Type":"ContainerStarted","Data":"b8f3e5b2811ca54e3cb67ba94caa41bf59c533c9d599f9d321b2b3e02d5b6d3d"} Oct 07 15:05:44 crc kubenswrapper[4959]: I1007 15:05:44.137890 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rphjb" podStartSLOduration=2.602991667 podStartE2EDuration="5.137869328s" podCreationTimestamp="2025-10-07 15:05:39 +0000 UTC" firstStartedPulling="2025-10-07 15:05:41.079216845 +0000 UTC m=+4823.162621170" lastFinishedPulling="2025-10-07 15:05:43.614094516 +0000 UTC m=+4825.697498831" observedRunningTime="2025-10-07 15:05:44.133696262 +0000 UTC m=+4826.217100607" watchObservedRunningTime="2025-10-07 15:05:44.137869328 +0000 UTC m=+4826.221273653" Oct 07 15:05:49 crc kubenswrapper[4959]: I1007 15:05:49.703158 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rphjb" Oct 07 15:05:49 crc kubenswrapper[4959]: I1007 15:05:49.703837 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rphjb" Oct 07 15:05:49 crc kubenswrapper[4959]: I1007 15:05:49.761305 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rphjb" Oct 07 15:05:50 crc kubenswrapper[4959]: I1007 15:05:50.229695 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rphjb" Oct 07 15:05:50 crc kubenswrapper[4959]: I1007 15:05:50.292818 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rphjb"] Oct 07 15:05:50 crc kubenswrapper[4959]: I1007 15:05:50.630242 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 15:05:50 crc kubenswrapper[4959]: I1007 15:05:50.630373 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 15:05:50 crc kubenswrapper[4959]: I1007 15:05:50.630454 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 15:05:50 crc kubenswrapper[4959]: I1007 15:05:50.631419 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e136903417b368464eb103f0eb291b140d45a4b3861749d2f8a77717fa3714ce"} pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 15:05:50 crc kubenswrapper[4959]: I1007 15:05:50.631494 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" containerID="cri-o://e136903417b368464eb103f0eb291b140d45a4b3861749d2f8a77717fa3714ce" gracePeriod=600 Oct 07 15:05:51 crc kubenswrapper[4959]: I1007 15:05:51.197279 4959 generic.go:334] "Generic (PLEG): container finished" podID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerID="e136903417b368464eb103f0eb291b140d45a4b3861749d2f8a77717fa3714ce" exitCode=0 Oct 07 15:05:51 crc kubenswrapper[4959]: I1007 15:05:51.197340 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerDied","Data":"e136903417b368464eb103f0eb291b140d45a4b3861749d2f8a77717fa3714ce"} Oct 07 15:05:51 crc kubenswrapper[4959]: I1007 15:05:51.198297 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerStarted","Data":"e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074"} Oct 07 15:05:51 crc kubenswrapper[4959]: I1007 15:05:51.198328 4959 scope.go:117] "RemoveContainer" containerID="9174af3525470c26554fbc1f790c7a2477048ca206968428f1ef56f61038e204" Oct 07 15:05:52 crc kubenswrapper[4959]: I1007 15:05:52.215528 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rphjb" podUID="2ea7afff-d8c0-4ecd-9a84-17049000c904" containerName="registry-server" containerID="cri-o://b8f3e5b2811ca54e3cb67ba94caa41bf59c533c9d599f9d321b2b3e02d5b6d3d" gracePeriod=2 Oct 07 15:05:52 crc kubenswrapper[4959]: I1007 15:05:52.896720 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rphjb" Oct 07 15:05:52 crc kubenswrapper[4959]: I1007 15:05:52.996592 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ea7afff-d8c0-4ecd-9a84-17049000c904-utilities\") pod \"2ea7afff-d8c0-4ecd-9a84-17049000c904\" (UID: \"2ea7afff-d8c0-4ecd-9a84-17049000c904\") " Oct 07 15:05:52 crc kubenswrapper[4959]: I1007 15:05:52.996769 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ea7afff-d8c0-4ecd-9a84-17049000c904-catalog-content\") pod \"2ea7afff-d8c0-4ecd-9a84-17049000c904\" (UID: \"2ea7afff-d8c0-4ecd-9a84-17049000c904\") " Oct 07 15:05:52 crc kubenswrapper[4959]: I1007 15:05:52.996895 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvrf5\" (UniqueName: \"kubernetes.io/projected/2ea7afff-d8c0-4ecd-9a84-17049000c904-kube-api-access-cvrf5\") pod \"2ea7afff-d8c0-4ecd-9a84-17049000c904\" (UID: \"2ea7afff-d8c0-4ecd-9a84-17049000c904\") " Oct 07 15:05:52 crc kubenswrapper[4959]: I1007 15:05:52.999374 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ea7afff-d8c0-4ecd-9a84-17049000c904-utilities" (OuterVolumeSpecName: "utilities") pod "2ea7afff-d8c0-4ecd-9a84-17049000c904" (UID: "2ea7afff-d8c0-4ecd-9a84-17049000c904"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:05:53 crc kubenswrapper[4959]: I1007 15:05:53.005774 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ea7afff-d8c0-4ecd-9a84-17049000c904-kube-api-access-cvrf5" (OuterVolumeSpecName: "kube-api-access-cvrf5") pod "2ea7afff-d8c0-4ecd-9a84-17049000c904" (UID: "2ea7afff-d8c0-4ecd-9a84-17049000c904"). InnerVolumeSpecName "kube-api-access-cvrf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 15:05:53 crc kubenswrapper[4959]: I1007 15:05:53.012605 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ea7afff-d8c0-4ecd-9a84-17049000c904-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2ea7afff-d8c0-4ecd-9a84-17049000c904" (UID: "2ea7afff-d8c0-4ecd-9a84-17049000c904"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:05:53 crc kubenswrapper[4959]: I1007 15:05:53.099910 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ea7afff-d8c0-4ecd-9a84-17049000c904-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 15:05:53 crc kubenswrapper[4959]: I1007 15:05:53.100001 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvrf5\" (UniqueName: \"kubernetes.io/projected/2ea7afff-d8c0-4ecd-9a84-17049000c904-kube-api-access-cvrf5\") on node \"crc\" DevicePath \"\"" Oct 07 15:05:53 crc kubenswrapper[4959]: I1007 15:05:53.100017 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ea7afff-d8c0-4ecd-9a84-17049000c904-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 15:05:53 crc kubenswrapper[4959]: I1007 15:05:53.230896 4959 generic.go:334] "Generic (PLEG): container finished" podID="2ea7afff-d8c0-4ecd-9a84-17049000c904" containerID="b8f3e5b2811ca54e3cb67ba94caa41bf59c533c9d599f9d321b2b3e02d5b6d3d" exitCode=0 Oct 07 15:05:53 crc kubenswrapper[4959]: I1007 15:05:53.231027 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rphjb" event={"ID":"2ea7afff-d8c0-4ecd-9a84-17049000c904","Type":"ContainerDied","Data":"b8f3e5b2811ca54e3cb67ba94caa41bf59c533c9d599f9d321b2b3e02d5b6d3d"} Oct 07 15:05:53 crc kubenswrapper[4959]: I1007 15:05:53.231115 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rphjb" event={"ID":"2ea7afff-d8c0-4ecd-9a84-17049000c904","Type":"ContainerDied","Data":"ff4867a9ea8753a8e62d7b8c5f1d9ff7f4bb09a9ab5a4dad124d5509ed32570e"} Oct 07 15:05:53 crc kubenswrapper[4959]: I1007 15:05:53.231143 4959 scope.go:117] "RemoveContainer" containerID="b8f3e5b2811ca54e3cb67ba94caa41bf59c533c9d599f9d321b2b3e02d5b6d3d" Oct 07 15:05:53 crc kubenswrapper[4959]: I1007 15:05:53.231320 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rphjb" Oct 07 15:05:53 crc kubenswrapper[4959]: I1007 15:05:53.275991 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rphjb"] Oct 07 15:05:53 crc kubenswrapper[4959]: I1007 15:05:53.278842 4959 scope.go:117] "RemoveContainer" containerID="c3003a57ba13d8aa519ecc51f8c5260b12ef8c8a88876a2f14ceccc97bed6ca7" Oct 07 15:05:53 crc kubenswrapper[4959]: I1007 15:05:53.285841 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rphjb"] Oct 07 15:05:53 crc kubenswrapper[4959]: I1007 15:05:53.997372 4959 scope.go:117] "RemoveContainer" containerID="5fcc03a706183c519c10f3d7d49240d31b3c0319189ab9b57a20421158bfdd48" Oct 07 15:05:54 crc kubenswrapper[4959]: I1007 15:05:54.056608 4959 scope.go:117] "RemoveContainer" containerID="b8f3e5b2811ca54e3cb67ba94caa41bf59c533c9d599f9d321b2b3e02d5b6d3d" Oct 07 15:05:54 crc kubenswrapper[4959]: E1007 15:05:54.057361 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8f3e5b2811ca54e3cb67ba94caa41bf59c533c9d599f9d321b2b3e02d5b6d3d\": container with ID starting with b8f3e5b2811ca54e3cb67ba94caa41bf59c533c9d599f9d321b2b3e02d5b6d3d not found: ID does not exist" containerID="b8f3e5b2811ca54e3cb67ba94caa41bf59c533c9d599f9d321b2b3e02d5b6d3d" Oct 07 15:05:54 crc kubenswrapper[4959]: I1007 15:05:54.057412 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8f3e5b2811ca54e3cb67ba94caa41bf59c533c9d599f9d321b2b3e02d5b6d3d"} err="failed to get container status \"b8f3e5b2811ca54e3cb67ba94caa41bf59c533c9d599f9d321b2b3e02d5b6d3d\": rpc error: code = NotFound desc = could not find container \"b8f3e5b2811ca54e3cb67ba94caa41bf59c533c9d599f9d321b2b3e02d5b6d3d\": container with ID starting with b8f3e5b2811ca54e3cb67ba94caa41bf59c533c9d599f9d321b2b3e02d5b6d3d not found: ID does not exist" Oct 07 15:05:54 crc kubenswrapper[4959]: I1007 15:05:54.057439 4959 scope.go:117] "RemoveContainer" containerID="c3003a57ba13d8aa519ecc51f8c5260b12ef8c8a88876a2f14ceccc97bed6ca7" Oct 07 15:05:54 crc kubenswrapper[4959]: E1007 15:05:54.058019 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3003a57ba13d8aa519ecc51f8c5260b12ef8c8a88876a2f14ceccc97bed6ca7\": container with ID starting with c3003a57ba13d8aa519ecc51f8c5260b12ef8c8a88876a2f14ceccc97bed6ca7 not found: ID does not exist" containerID="c3003a57ba13d8aa519ecc51f8c5260b12ef8c8a88876a2f14ceccc97bed6ca7" Oct 07 15:05:54 crc kubenswrapper[4959]: I1007 15:05:54.058047 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3003a57ba13d8aa519ecc51f8c5260b12ef8c8a88876a2f14ceccc97bed6ca7"} err="failed to get container status \"c3003a57ba13d8aa519ecc51f8c5260b12ef8c8a88876a2f14ceccc97bed6ca7\": rpc error: code = NotFound desc = could not find container \"c3003a57ba13d8aa519ecc51f8c5260b12ef8c8a88876a2f14ceccc97bed6ca7\": container with ID starting with c3003a57ba13d8aa519ecc51f8c5260b12ef8c8a88876a2f14ceccc97bed6ca7 not found: ID does not exist" Oct 07 15:05:54 crc kubenswrapper[4959]: I1007 15:05:54.058082 4959 scope.go:117] "RemoveContainer" containerID="5fcc03a706183c519c10f3d7d49240d31b3c0319189ab9b57a20421158bfdd48" Oct 07 15:05:54 crc kubenswrapper[4959]: E1007 15:05:54.058621 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fcc03a706183c519c10f3d7d49240d31b3c0319189ab9b57a20421158bfdd48\": container with ID starting with 5fcc03a706183c519c10f3d7d49240d31b3c0319189ab9b57a20421158bfdd48 not found: ID does not exist" containerID="5fcc03a706183c519c10f3d7d49240d31b3c0319189ab9b57a20421158bfdd48" Oct 07 15:05:54 crc kubenswrapper[4959]: I1007 15:05:54.058657 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fcc03a706183c519c10f3d7d49240d31b3c0319189ab9b57a20421158bfdd48"} err="failed to get container status \"5fcc03a706183c519c10f3d7d49240d31b3c0319189ab9b57a20421158bfdd48\": rpc error: code = NotFound desc = could not find container \"5fcc03a706183c519c10f3d7d49240d31b3c0319189ab9b57a20421158bfdd48\": container with ID starting with 5fcc03a706183c519c10f3d7d49240d31b3c0319189ab9b57a20421158bfdd48 not found: ID does not exist" Oct 07 15:05:54 crc kubenswrapper[4959]: I1007 15:05:54.664839 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ea7afff-d8c0-4ecd-9a84-17049000c904" path="/var/lib/kubelet/pods/2ea7afff-d8c0-4ecd-9a84-17049000c904/volumes" Oct 07 15:07:03 crc kubenswrapper[4959]: I1007 15:07:03.745886 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fkjqb"] Oct 07 15:07:03 crc kubenswrapper[4959]: E1007 15:07:03.747047 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ea7afff-d8c0-4ecd-9a84-17049000c904" containerName="registry-server" Oct 07 15:07:03 crc kubenswrapper[4959]: I1007 15:07:03.747064 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ea7afff-d8c0-4ecd-9a84-17049000c904" containerName="registry-server" Oct 07 15:07:03 crc kubenswrapper[4959]: E1007 15:07:03.747084 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ea7afff-d8c0-4ecd-9a84-17049000c904" containerName="extract-utilities" Oct 07 15:07:03 crc kubenswrapper[4959]: I1007 15:07:03.747094 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ea7afff-d8c0-4ecd-9a84-17049000c904" containerName="extract-utilities" Oct 07 15:07:03 crc kubenswrapper[4959]: E1007 15:07:03.747128 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ea7afff-d8c0-4ecd-9a84-17049000c904" containerName="extract-content" Oct 07 15:07:03 crc kubenswrapper[4959]: I1007 15:07:03.747137 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ea7afff-d8c0-4ecd-9a84-17049000c904" containerName="extract-content" Oct 07 15:07:03 crc kubenswrapper[4959]: I1007 15:07:03.747429 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ea7afff-d8c0-4ecd-9a84-17049000c904" containerName="registry-server" Oct 07 15:07:03 crc kubenswrapper[4959]: I1007 15:07:03.749257 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fkjqb" Oct 07 15:07:03 crc kubenswrapper[4959]: I1007 15:07:03.767447 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fkjqb"] Oct 07 15:07:03 crc kubenswrapper[4959]: I1007 15:07:03.856179 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58e3bc63-3729-43bb-b8df-e0a9e45c2ac6-utilities\") pod \"community-operators-fkjqb\" (UID: \"58e3bc63-3729-43bb-b8df-e0a9e45c2ac6\") " pod="openshift-marketplace/community-operators-fkjqb" Oct 07 15:07:03 crc kubenswrapper[4959]: I1007 15:07:03.856489 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44fm9\" (UniqueName: \"kubernetes.io/projected/58e3bc63-3729-43bb-b8df-e0a9e45c2ac6-kube-api-access-44fm9\") pod \"community-operators-fkjqb\" (UID: \"58e3bc63-3729-43bb-b8df-e0a9e45c2ac6\") " pod="openshift-marketplace/community-operators-fkjqb" Oct 07 15:07:03 crc kubenswrapper[4959]: I1007 15:07:03.856580 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58e3bc63-3729-43bb-b8df-e0a9e45c2ac6-catalog-content\") pod \"community-operators-fkjqb\" (UID: \"58e3bc63-3729-43bb-b8df-e0a9e45c2ac6\") " pod="openshift-marketplace/community-operators-fkjqb" Oct 07 15:07:03 crc kubenswrapper[4959]: I1007 15:07:03.958871 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58e3bc63-3729-43bb-b8df-e0a9e45c2ac6-utilities\") pod \"community-operators-fkjqb\" (UID: \"58e3bc63-3729-43bb-b8df-e0a9e45c2ac6\") " pod="openshift-marketplace/community-operators-fkjqb" Oct 07 15:07:03 crc kubenswrapper[4959]: I1007 15:07:03.959010 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44fm9\" (UniqueName: \"kubernetes.io/projected/58e3bc63-3729-43bb-b8df-e0a9e45c2ac6-kube-api-access-44fm9\") pod \"community-operators-fkjqb\" (UID: \"58e3bc63-3729-43bb-b8df-e0a9e45c2ac6\") " pod="openshift-marketplace/community-operators-fkjqb" Oct 07 15:07:03 crc kubenswrapper[4959]: I1007 15:07:03.959044 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58e3bc63-3729-43bb-b8df-e0a9e45c2ac6-catalog-content\") pod \"community-operators-fkjqb\" (UID: \"58e3bc63-3729-43bb-b8df-e0a9e45c2ac6\") " pod="openshift-marketplace/community-operators-fkjqb" Oct 07 15:07:03 crc kubenswrapper[4959]: I1007 15:07:03.959573 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58e3bc63-3729-43bb-b8df-e0a9e45c2ac6-catalog-content\") pod \"community-operators-fkjqb\" (UID: \"58e3bc63-3729-43bb-b8df-e0a9e45c2ac6\") " pod="openshift-marketplace/community-operators-fkjqb" Oct 07 15:07:03 crc kubenswrapper[4959]: I1007 15:07:03.959833 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58e3bc63-3729-43bb-b8df-e0a9e45c2ac6-utilities\") pod \"community-operators-fkjqb\" (UID: \"58e3bc63-3729-43bb-b8df-e0a9e45c2ac6\") " pod="openshift-marketplace/community-operators-fkjqb" Oct 07 15:07:03 crc kubenswrapper[4959]: I1007 15:07:03.990726 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44fm9\" (UniqueName: \"kubernetes.io/projected/58e3bc63-3729-43bb-b8df-e0a9e45c2ac6-kube-api-access-44fm9\") pod \"community-operators-fkjqb\" (UID: \"58e3bc63-3729-43bb-b8df-e0a9e45c2ac6\") " pod="openshift-marketplace/community-operators-fkjqb" Oct 07 15:07:04 crc kubenswrapper[4959]: I1007 15:07:04.086268 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fkjqb" Oct 07 15:07:04 crc kubenswrapper[4959]: I1007 15:07:04.672038 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fkjqb"] Oct 07 15:07:04 crc kubenswrapper[4959]: I1007 15:07:04.950262 4959 generic.go:334] "Generic (PLEG): container finished" podID="58e3bc63-3729-43bb-b8df-e0a9e45c2ac6" containerID="9935b3ee4fbc508b293a06d2ec8e8aed756007dc2b49a80ee1a39475e677eaae" exitCode=0 Oct 07 15:07:04 crc kubenswrapper[4959]: I1007 15:07:04.950306 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fkjqb" event={"ID":"58e3bc63-3729-43bb-b8df-e0a9e45c2ac6","Type":"ContainerDied","Data":"9935b3ee4fbc508b293a06d2ec8e8aed756007dc2b49a80ee1a39475e677eaae"} Oct 07 15:07:04 crc kubenswrapper[4959]: I1007 15:07:04.950353 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fkjqb" event={"ID":"58e3bc63-3729-43bb-b8df-e0a9e45c2ac6","Type":"ContainerStarted","Data":"c81b799a38fb20c225427974984911b318ef512b7f6f08326a094f4d6b50e508"} Oct 07 15:07:06 crc kubenswrapper[4959]: I1007 15:07:06.971560 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fkjqb" event={"ID":"58e3bc63-3729-43bb-b8df-e0a9e45c2ac6","Type":"ContainerStarted","Data":"1c0f6f07917cab23ccc5bfc0a9198503291a4f2733c7c90c35c391edc0413b06"} Oct 07 15:07:07 crc kubenswrapper[4959]: I1007 15:07:07.982419 4959 generic.go:334] "Generic (PLEG): container finished" podID="58e3bc63-3729-43bb-b8df-e0a9e45c2ac6" containerID="1c0f6f07917cab23ccc5bfc0a9198503291a4f2733c7c90c35c391edc0413b06" exitCode=0 Oct 07 15:07:07 crc kubenswrapper[4959]: I1007 15:07:07.982528 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fkjqb" event={"ID":"58e3bc63-3729-43bb-b8df-e0a9e45c2ac6","Type":"ContainerDied","Data":"1c0f6f07917cab23ccc5bfc0a9198503291a4f2733c7c90c35c391edc0413b06"} Oct 07 15:07:08 crc kubenswrapper[4959]: I1007 15:07:08.994491 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fkjqb" event={"ID":"58e3bc63-3729-43bb-b8df-e0a9e45c2ac6","Type":"ContainerStarted","Data":"53d351c448bc07d64e81bcd68f841e1bc7d2c19bd046a517da6c5527507a7768"} Oct 07 15:07:09 crc kubenswrapper[4959]: I1007 15:07:09.026598 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fkjqb" podStartSLOduration=2.579256391 podStartE2EDuration="6.026577585s" podCreationTimestamp="2025-10-07 15:07:03 +0000 UTC" firstStartedPulling="2025-10-07 15:07:04.952864935 +0000 UTC m=+4907.036269260" lastFinishedPulling="2025-10-07 15:07:08.400186129 +0000 UTC m=+4910.483590454" observedRunningTime="2025-10-07 15:07:09.019927346 +0000 UTC m=+4911.103331671" watchObservedRunningTime="2025-10-07 15:07:09.026577585 +0000 UTC m=+4911.109981910" Oct 07 15:07:14 crc kubenswrapper[4959]: I1007 15:07:14.086906 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fkjqb" Oct 07 15:07:14 crc kubenswrapper[4959]: I1007 15:07:14.090733 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fkjqb" Oct 07 15:07:14 crc kubenswrapper[4959]: I1007 15:07:14.145633 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fkjqb" Oct 07 15:07:15 crc kubenswrapper[4959]: I1007 15:07:15.096082 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fkjqb" Oct 07 15:07:15 crc kubenswrapper[4959]: I1007 15:07:15.150736 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fkjqb"] Oct 07 15:07:17 crc kubenswrapper[4959]: I1007 15:07:17.066503 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fkjqb" podUID="58e3bc63-3729-43bb-b8df-e0a9e45c2ac6" containerName="registry-server" containerID="cri-o://53d351c448bc07d64e81bcd68f841e1bc7d2c19bd046a517da6c5527507a7768" gracePeriod=2 Oct 07 15:07:17 crc kubenswrapper[4959]: I1007 15:07:17.697987 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fkjqb" Oct 07 15:07:17 crc kubenswrapper[4959]: I1007 15:07:17.772638 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58e3bc63-3729-43bb-b8df-e0a9e45c2ac6-catalog-content\") pod \"58e3bc63-3729-43bb-b8df-e0a9e45c2ac6\" (UID: \"58e3bc63-3729-43bb-b8df-e0a9e45c2ac6\") " Oct 07 15:07:17 crc kubenswrapper[4959]: I1007 15:07:17.772790 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58e3bc63-3729-43bb-b8df-e0a9e45c2ac6-utilities\") pod \"58e3bc63-3729-43bb-b8df-e0a9e45c2ac6\" (UID: \"58e3bc63-3729-43bb-b8df-e0a9e45c2ac6\") " Oct 07 15:07:17 crc kubenswrapper[4959]: I1007 15:07:17.772998 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44fm9\" (UniqueName: \"kubernetes.io/projected/58e3bc63-3729-43bb-b8df-e0a9e45c2ac6-kube-api-access-44fm9\") pod \"58e3bc63-3729-43bb-b8df-e0a9e45c2ac6\" (UID: \"58e3bc63-3729-43bb-b8df-e0a9e45c2ac6\") " Oct 07 15:07:17 crc kubenswrapper[4959]: I1007 15:07:17.773880 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58e3bc63-3729-43bb-b8df-e0a9e45c2ac6-utilities" (OuterVolumeSpecName: "utilities") pod "58e3bc63-3729-43bb-b8df-e0a9e45c2ac6" (UID: "58e3bc63-3729-43bb-b8df-e0a9e45c2ac6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:07:17 crc kubenswrapper[4959]: I1007 15:07:17.779288 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58e3bc63-3729-43bb-b8df-e0a9e45c2ac6-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 15:07:17 crc kubenswrapper[4959]: I1007 15:07:17.786460 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58e3bc63-3729-43bb-b8df-e0a9e45c2ac6-kube-api-access-44fm9" (OuterVolumeSpecName: "kube-api-access-44fm9") pod "58e3bc63-3729-43bb-b8df-e0a9e45c2ac6" (UID: "58e3bc63-3729-43bb-b8df-e0a9e45c2ac6"). InnerVolumeSpecName "kube-api-access-44fm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 15:07:17 crc kubenswrapper[4959]: I1007 15:07:17.831306 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58e3bc63-3729-43bb-b8df-e0a9e45c2ac6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "58e3bc63-3729-43bb-b8df-e0a9e45c2ac6" (UID: "58e3bc63-3729-43bb-b8df-e0a9e45c2ac6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:07:17 crc kubenswrapper[4959]: I1007 15:07:17.882780 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58e3bc63-3729-43bb-b8df-e0a9e45c2ac6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 15:07:17 crc kubenswrapper[4959]: I1007 15:07:17.882842 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44fm9\" (UniqueName: \"kubernetes.io/projected/58e3bc63-3729-43bb-b8df-e0a9e45c2ac6-kube-api-access-44fm9\") on node \"crc\" DevicePath \"\"" Oct 07 15:07:18 crc kubenswrapper[4959]: I1007 15:07:18.082490 4959 generic.go:334] "Generic (PLEG): container finished" podID="58e3bc63-3729-43bb-b8df-e0a9e45c2ac6" containerID="53d351c448bc07d64e81bcd68f841e1bc7d2c19bd046a517da6c5527507a7768" exitCode=0 Oct 07 15:07:18 crc kubenswrapper[4959]: I1007 15:07:18.082553 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fkjqb" event={"ID":"58e3bc63-3729-43bb-b8df-e0a9e45c2ac6","Type":"ContainerDied","Data":"53d351c448bc07d64e81bcd68f841e1bc7d2c19bd046a517da6c5527507a7768"} Oct 07 15:07:18 crc kubenswrapper[4959]: I1007 15:07:18.082568 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fkjqb" Oct 07 15:07:18 crc kubenswrapper[4959]: I1007 15:07:18.082600 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fkjqb" event={"ID":"58e3bc63-3729-43bb-b8df-e0a9e45c2ac6","Type":"ContainerDied","Data":"c81b799a38fb20c225427974984911b318ef512b7f6f08326a094f4d6b50e508"} Oct 07 15:07:18 crc kubenswrapper[4959]: I1007 15:07:18.082625 4959 scope.go:117] "RemoveContainer" containerID="53d351c448bc07d64e81bcd68f841e1bc7d2c19bd046a517da6c5527507a7768" Oct 07 15:07:18 crc kubenswrapper[4959]: I1007 15:07:18.119850 4959 scope.go:117] "RemoveContainer" containerID="1c0f6f07917cab23ccc5bfc0a9198503291a4f2733c7c90c35c391edc0413b06" Oct 07 15:07:18 crc kubenswrapper[4959]: I1007 15:07:18.121924 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fkjqb"] Oct 07 15:07:18 crc kubenswrapper[4959]: I1007 15:07:18.146304 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fkjqb"] Oct 07 15:07:18 crc kubenswrapper[4959]: I1007 15:07:18.152282 4959 scope.go:117] "RemoveContainer" containerID="9935b3ee4fbc508b293a06d2ec8e8aed756007dc2b49a80ee1a39475e677eaae" Oct 07 15:07:18 crc kubenswrapper[4959]: I1007 15:07:18.198526 4959 scope.go:117] "RemoveContainer" containerID="53d351c448bc07d64e81bcd68f841e1bc7d2c19bd046a517da6c5527507a7768" Oct 07 15:07:18 crc kubenswrapper[4959]: E1007 15:07:18.199347 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53d351c448bc07d64e81bcd68f841e1bc7d2c19bd046a517da6c5527507a7768\": container with ID starting with 53d351c448bc07d64e81bcd68f841e1bc7d2c19bd046a517da6c5527507a7768 not found: ID does not exist" containerID="53d351c448bc07d64e81bcd68f841e1bc7d2c19bd046a517da6c5527507a7768" Oct 07 15:07:18 crc kubenswrapper[4959]: I1007 15:07:18.199446 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53d351c448bc07d64e81bcd68f841e1bc7d2c19bd046a517da6c5527507a7768"} err="failed to get container status \"53d351c448bc07d64e81bcd68f841e1bc7d2c19bd046a517da6c5527507a7768\": rpc error: code = NotFound desc = could not find container \"53d351c448bc07d64e81bcd68f841e1bc7d2c19bd046a517da6c5527507a7768\": container with ID starting with 53d351c448bc07d64e81bcd68f841e1bc7d2c19bd046a517da6c5527507a7768 not found: ID does not exist" Oct 07 15:07:18 crc kubenswrapper[4959]: I1007 15:07:18.199497 4959 scope.go:117] "RemoveContainer" containerID="1c0f6f07917cab23ccc5bfc0a9198503291a4f2733c7c90c35c391edc0413b06" Oct 07 15:07:18 crc kubenswrapper[4959]: E1007 15:07:18.200279 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c0f6f07917cab23ccc5bfc0a9198503291a4f2733c7c90c35c391edc0413b06\": container with ID starting with 1c0f6f07917cab23ccc5bfc0a9198503291a4f2733c7c90c35c391edc0413b06 not found: ID does not exist" containerID="1c0f6f07917cab23ccc5bfc0a9198503291a4f2733c7c90c35c391edc0413b06" Oct 07 15:07:18 crc kubenswrapper[4959]: I1007 15:07:18.200362 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c0f6f07917cab23ccc5bfc0a9198503291a4f2733c7c90c35c391edc0413b06"} err="failed to get container status \"1c0f6f07917cab23ccc5bfc0a9198503291a4f2733c7c90c35c391edc0413b06\": rpc error: code = NotFound desc = could not find container \"1c0f6f07917cab23ccc5bfc0a9198503291a4f2733c7c90c35c391edc0413b06\": container with ID starting with 1c0f6f07917cab23ccc5bfc0a9198503291a4f2733c7c90c35c391edc0413b06 not found: ID does not exist" Oct 07 15:07:18 crc kubenswrapper[4959]: I1007 15:07:18.200437 4959 scope.go:117] "RemoveContainer" containerID="9935b3ee4fbc508b293a06d2ec8e8aed756007dc2b49a80ee1a39475e677eaae" Oct 07 15:07:18 crc kubenswrapper[4959]: E1007 15:07:18.201250 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9935b3ee4fbc508b293a06d2ec8e8aed756007dc2b49a80ee1a39475e677eaae\": container with ID starting with 9935b3ee4fbc508b293a06d2ec8e8aed756007dc2b49a80ee1a39475e677eaae not found: ID does not exist" containerID="9935b3ee4fbc508b293a06d2ec8e8aed756007dc2b49a80ee1a39475e677eaae" Oct 07 15:07:18 crc kubenswrapper[4959]: I1007 15:07:18.201286 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9935b3ee4fbc508b293a06d2ec8e8aed756007dc2b49a80ee1a39475e677eaae"} err="failed to get container status \"9935b3ee4fbc508b293a06d2ec8e8aed756007dc2b49a80ee1a39475e677eaae\": rpc error: code = NotFound desc = could not find container \"9935b3ee4fbc508b293a06d2ec8e8aed756007dc2b49a80ee1a39475e677eaae\": container with ID starting with 9935b3ee4fbc508b293a06d2ec8e8aed756007dc2b49a80ee1a39475e677eaae not found: ID does not exist" Oct 07 15:07:18 crc kubenswrapper[4959]: I1007 15:07:18.668771 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58e3bc63-3729-43bb-b8df-e0a9e45c2ac6" path="/var/lib/kubelet/pods/58e3bc63-3729-43bb-b8df-e0a9e45c2ac6/volumes" Oct 07 15:08:20 crc kubenswrapper[4959]: I1007 15:08:20.630928 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 15:08:20 crc kubenswrapper[4959]: I1007 15:08:20.631705 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 15:08:50 crc kubenswrapper[4959]: I1007 15:08:50.630000 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 15:08:50 crc kubenswrapper[4959]: I1007 15:08:50.630795 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 15:09:20 crc kubenswrapper[4959]: I1007 15:09:20.630498 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 15:09:20 crc kubenswrapper[4959]: I1007 15:09:20.631225 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 15:09:20 crc kubenswrapper[4959]: I1007 15:09:20.631282 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 15:09:20 crc kubenswrapper[4959]: I1007 15:09:20.632306 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074"} pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 15:09:20 crc kubenswrapper[4959]: I1007 15:09:20.632373 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" containerID="cri-o://e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074" gracePeriod=600 Oct 07 15:09:21 crc kubenswrapper[4959]: I1007 15:09:21.262510 4959 generic.go:334] "Generic (PLEG): container finished" podID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerID="e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074" exitCode=0 Oct 07 15:09:21 crc kubenswrapper[4959]: I1007 15:09:21.262614 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerDied","Data":"e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074"} Oct 07 15:09:21 crc kubenswrapper[4959]: I1007 15:09:21.263053 4959 scope.go:117] "RemoveContainer" containerID="e136903417b368464eb103f0eb291b140d45a4b3861749d2f8a77717fa3714ce" Oct 07 15:09:21 crc kubenswrapper[4959]: E1007 15:09:21.293967 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:09:22 crc kubenswrapper[4959]: I1007 15:09:22.277730 4959 scope.go:117] "RemoveContainer" containerID="e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074" Oct 07 15:09:22 crc kubenswrapper[4959]: E1007 15:09:22.278058 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:09:36 crc kubenswrapper[4959]: I1007 15:09:36.654300 4959 scope.go:117] "RemoveContainer" containerID="e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074" Oct 07 15:09:36 crc kubenswrapper[4959]: E1007 15:09:36.655381 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:09:49 crc kubenswrapper[4959]: I1007 15:09:49.655694 4959 scope.go:117] "RemoveContainer" containerID="e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074" Oct 07 15:09:49 crc kubenswrapper[4959]: E1007 15:09:49.656616 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:10:01 crc kubenswrapper[4959]: I1007 15:10:01.654149 4959 scope.go:117] "RemoveContainer" containerID="e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074" Oct 07 15:10:01 crc kubenswrapper[4959]: E1007 15:10:01.655075 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:10:13 crc kubenswrapper[4959]: I1007 15:10:13.653887 4959 scope.go:117] "RemoveContainer" containerID="e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074" Oct 07 15:10:13 crc kubenswrapper[4959]: E1007 15:10:13.654721 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:10:28 crc kubenswrapper[4959]: I1007 15:10:28.662471 4959 scope.go:117] "RemoveContainer" containerID="e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074" Oct 07 15:10:28 crc kubenswrapper[4959]: E1007 15:10:28.663690 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:10:39 crc kubenswrapper[4959]: I1007 15:10:39.654774 4959 scope.go:117] "RemoveContainer" containerID="e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074" Oct 07 15:10:39 crc kubenswrapper[4959]: E1007 15:10:39.655986 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:10:54 crc kubenswrapper[4959]: I1007 15:10:54.654200 4959 scope.go:117] "RemoveContainer" containerID="e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074" Oct 07 15:10:54 crc kubenswrapper[4959]: E1007 15:10:54.655771 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:11:06 crc kubenswrapper[4959]: I1007 15:11:06.653999 4959 scope.go:117] "RemoveContainer" containerID="e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074" Oct 07 15:11:06 crc kubenswrapper[4959]: E1007 15:11:06.654818 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:11:17 crc kubenswrapper[4959]: I1007 15:11:17.653886 4959 scope.go:117] "RemoveContainer" containerID="e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074" Oct 07 15:11:17 crc kubenswrapper[4959]: E1007 15:11:17.654912 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:11:30 crc kubenswrapper[4959]: I1007 15:11:30.653802 4959 scope.go:117] "RemoveContainer" containerID="e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074" Oct 07 15:11:30 crc kubenswrapper[4959]: E1007 15:11:30.656077 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:11:44 crc kubenswrapper[4959]: I1007 15:11:44.654052 4959 scope.go:117] "RemoveContainer" containerID="e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074" Oct 07 15:11:44 crc kubenswrapper[4959]: E1007 15:11:44.654911 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:11:57 crc kubenswrapper[4959]: I1007 15:11:57.654623 4959 scope.go:117] "RemoveContainer" containerID="e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074" Oct 07 15:11:57 crc kubenswrapper[4959]: E1007 15:11:57.655617 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:12:12 crc kubenswrapper[4959]: I1007 15:12:12.654012 4959 scope.go:117] "RemoveContainer" containerID="e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074" Oct 07 15:12:12 crc kubenswrapper[4959]: E1007 15:12:12.654927 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:12:26 crc kubenswrapper[4959]: I1007 15:12:26.653498 4959 scope.go:117] "RemoveContainer" containerID="e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074" Oct 07 15:12:26 crc kubenswrapper[4959]: E1007 15:12:26.654432 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:12:38 crc kubenswrapper[4959]: I1007 15:12:38.660536 4959 scope.go:117] "RemoveContainer" containerID="e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074" Oct 07 15:12:38 crc kubenswrapper[4959]: E1007 15:12:38.661341 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:12:39 crc kubenswrapper[4959]: I1007 15:12:39.563352 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-txh4q"] Oct 07 15:12:39 crc kubenswrapper[4959]: E1007 15:12:39.565088 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58e3bc63-3729-43bb-b8df-e0a9e45c2ac6" containerName="registry-server" Oct 07 15:12:39 crc kubenswrapper[4959]: I1007 15:12:39.565114 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="58e3bc63-3729-43bb-b8df-e0a9e45c2ac6" containerName="registry-server" Oct 07 15:12:39 crc kubenswrapper[4959]: E1007 15:12:39.565131 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58e3bc63-3729-43bb-b8df-e0a9e45c2ac6" containerName="extract-utilities" Oct 07 15:12:39 crc kubenswrapper[4959]: I1007 15:12:39.565137 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="58e3bc63-3729-43bb-b8df-e0a9e45c2ac6" containerName="extract-utilities" Oct 07 15:12:39 crc kubenswrapper[4959]: E1007 15:12:39.565162 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58e3bc63-3729-43bb-b8df-e0a9e45c2ac6" containerName="extract-content" Oct 07 15:12:39 crc kubenswrapper[4959]: I1007 15:12:39.565170 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="58e3bc63-3729-43bb-b8df-e0a9e45c2ac6" containerName="extract-content" Oct 07 15:12:39 crc kubenswrapper[4959]: I1007 15:12:39.565377 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="58e3bc63-3729-43bb-b8df-e0a9e45c2ac6" containerName="registry-server" Oct 07 15:12:39 crc kubenswrapper[4959]: I1007 15:12:39.567309 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-txh4q" Oct 07 15:12:39 crc kubenswrapper[4959]: I1007 15:12:39.583508 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-txh4q"] Oct 07 15:12:39 crc kubenswrapper[4959]: I1007 15:12:39.597653 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43c37957-2e0c-44bb-9072-b5bcb256d772-catalog-content\") pod \"certified-operators-txh4q\" (UID: \"43c37957-2e0c-44bb-9072-b5bcb256d772\") " pod="openshift-marketplace/certified-operators-txh4q" Oct 07 15:12:39 crc kubenswrapper[4959]: I1007 15:12:39.597754 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43c37957-2e0c-44bb-9072-b5bcb256d772-utilities\") pod \"certified-operators-txh4q\" (UID: \"43c37957-2e0c-44bb-9072-b5bcb256d772\") " pod="openshift-marketplace/certified-operators-txh4q" Oct 07 15:12:39 crc kubenswrapper[4959]: I1007 15:12:39.597828 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvfg7\" (UniqueName: \"kubernetes.io/projected/43c37957-2e0c-44bb-9072-b5bcb256d772-kube-api-access-xvfg7\") pod \"certified-operators-txh4q\" (UID: \"43c37957-2e0c-44bb-9072-b5bcb256d772\") " pod="openshift-marketplace/certified-operators-txh4q" Oct 07 15:12:39 crc kubenswrapper[4959]: I1007 15:12:39.699918 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvfg7\" (UniqueName: \"kubernetes.io/projected/43c37957-2e0c-44bb-9072-b5bcb256d772-kube-api-access-xvfg7\") pod \"certified-operators-txh4q\" (UID: \"43c37957-2e0c-44bb-9072-b5bcb256d772\") " pod="openshift-marketplace/certified-operators-txh4q" Oct 07 15:12:39 crc kubenswrapper[4959]: I1007 15:12:39.700270 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43c37957-2e0c-44bb-9072-b5bcb256d772-catalog-content\") pod \"certified-operators-txh4q\" (UID: \"43c37957-2e0c-44bb-9072-b5bcb256d772\") " pod="openshift-marketplace/certified-operators-txh4q" Oct 07 15:12:39 crc kubenswrapper[4959]: I1007 15:12:39.700328 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43c37957-2e0c-44bb-9072-b5bcb256d772-utilities\") pod \"certified-operators-txh4q\" (UID: \"43c37957-2e0c-44bb-9072-b5bcb256d772\") " pod="openshift-marketplace/certified-operators-txh4q" Oct 07 15:12:39 crc kubenswrapper[4959]: I1007 15:12:39.701707 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43c37957-2e0c-44bb-9072-b5bcb256d772-utilities\") pod \"certified-operators-txh4q\" (UID: \"43c37957-2e0c-44bb-9072-b5bcb256d772\") " pod="openshift-marketplace/certified-operators-txh4q" Oct 07 15:12:39 crc kubenswrapper[4959]: I1007 15:12:39.701774 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43c37957-2e0c-44bb-9072-b5bcb256d772-catalog-content\") pod \"certified-operators-txh4q\" (UID: \"43c37957-2e0c-44bb-9072-b5bcb256d772\") " pod="openshift-marketplace/certified-operators-txh4q" Oct 07 15:12:39 crc kubenswrapper[4959]: I1007 15:12:39.724694 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvfg7\" (UniqueName: \"kubernetes.io/projected/43c37957-2e0c-44bb-9072-b5bcb256d772-kube-api-access-xvfg7\") pod \"certified-operators-txh4q\" (UID: \"43c37957-2e0c-44bb-9072-b5bcb256d772\") " pod="openshift-marketplace/certified-operators-txh4q" Oct 07 15:12:39 crc kubenswrapper[4959]: I1007 15:12:39.893234 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-txh4q" Oct 07 15:12:40 crc kubenswrapper[4959]: I1007 15:12:40.614771 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-txh4q"] Oct 07 15:12:41 crc kubenswrapper[4959]: I1007 15:12:41.276890 4959 generic.go:334] "Generic (PLEG): container finished" podID="43c37957-2e0c-44bb-9072-b5bcb256d772" containerID="d37ecb903482296123abb2583c8d4a87dbc9cad64b64a3fbbc3d729348a4b2e5" exitCode=0 Oct 07 15:12:41 crc kubenswrapper[4959]: I1007 15:12:41.277019 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-txh4q" event={"ID":"43c37957-2e0c-44bb-9072-b5bcb256d772","Type":"ContainerDied","Data":"d37ecb903482296123abb2583c8d4a87dbc9cad64b64a3fbbc3d729348a4b2e5"} Oct 07 15:12:41 crc kubenswrapper[4959]: I1007 15:12:41.277881 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-txh4q" event={"ID":"43c37957-2e0c-44bb-9072-b5bcb256d772","Type":"ContainerStarted","Data":"c9e53230e046485637f40b705560bebab5c6135be7168ddf416d276f407ecdd5"} Oct 07 15:12:41 crc kubenswrapper[4959]: I1007 15:12:41.280408 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 15:12:44 crc kubenswrapper[4959]: I1007 15:12:44.311903 4959 generic.go:334] "Generic (PLEG): container finished" podID="43c37957-2e0c-44bb-9072-b5bcb256d772" containerID="13d489449f923aa540a8b4876a9b4f309be07caba8289eaa4471f4e7bc6c13c2" exitCode=0 Oct 07 15:12:44 crc kubenswrapper[4959]: I1007 15:12:44.311965 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-txh4q" event={"ID":"43c37957-2e0c-44bb-9072-b5bcb256d772","Type":"ContainerDied","Data":"13d489449f923aa540a8b4876a9b4f309be07caba8289eaa4471f4e7bc6c13c2"} Oct 07 15:12:48 crc kubenswrapper[4959]: I1007 15:12:48.351515 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-txh4q" event={"ID":"43c37957-2e0c-44bb-9072-b5bcb256d772","Type":"ContainerStarted","Data":"d2bc7480abc71b7a9703c7f6ecc0c12387197b5ac794b7244a382b4fef3e29fd"} Oct 07 15:12:49 crc kubenswrapper[4959]: I1007 15:12:49.655306 4959 scope.go:117] "RemoveContainer" containerID="e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074" Oct 07 15:12:49 crc kubenswrapper[4959]: E1007 15:12:49.655599 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:12:49 crc kubenswrapper[4959]: I1007 15:12:49.893760 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-txh4q" Oct 07 15:12:49 crc kubenswrapper[4959]: I1007 15:12:49.893814 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-txh4q" Oct 07 15:12:49 crc kubenswrapper[4959]: I1007 15:12:49.944612 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-txh4q" Oct 07 15:12:49 crc kubenswrapper[4959]: I1007 15:12:49.972211 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-txh4q" podStartSLOduration=5.488159439 podStartE2EDuration="10.9721939s" podCreationTimestamp="2025-10-07 15:12:39 +0000 UTC" firstStartedPulling="2025-10-07 15:12:41.280053493 +0000 UTC m=+5243.363457818" lastFinishedPulling="2025-10-07 15:12:46.764087954 +0000 UTC m=+5248.847492279" observedRunningTime="2025-10-07 15:12:48.377754957 +0000 UTC m=+5250.461159292" watchObservedRunningTime="2025-10-07 15:12:49.9721939 +0000 UTC m=+5252.055598225" Oct 07 15:12:59 crc kubenswrapper[4959]: I1007 15:12:59.954954 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-txh4q" Oct 07 15:13:00 crc kubenswrapper[4959]: I1007 15:13:00.026590 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-txh4q"] Oct 07 15:13:00 crc kubenswrapper[4959]: I1007 15:13:00.474490 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-txh4q" podUID="43c37957-2e0c-44bb-9072-b5bcb256d772" containerName="registry-server" containerID="cri-o://d2bc7480abc71b7a9703c7f6ecc0c12387197b5ac794b7244a382b4fef3e29fd" gracePeriod=2 Oct 07 15:13:01 crc kubenswrapper[4959]: I1007 15:13:01.131445 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-txh4q" Oct 07 15:13:01 crc kubenswrapper[4959]: I1007 15:13:01.193779 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43c37957-2e0c-44bb-9072-b5bcb256d772-utilities\") pod \"43c37957-2e0c-44bb-9072-b5bcb256d772\" (UID: \"43c37957-2e0c-44bb-9072-b5bcb256d772\") " Oct 07 15:13:01 crc kubenswrapper[4959]: I1007 15:13:01.196944 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43c37957-2e0c-44bb-9072-b5bcb256d772-utilities" (OuterVolumeSpecName: "utilities") pod "43c37957-2e0c-44bb-9072-b5bcb256d772" (UID: "43c37957-2e0c-44bb-9072-b5bcb256d772"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:13:01 crc kubenswrapper[4959]: I1007 15:13:01.296454 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43c37957-2e0c-44bb-9072-b5bcb256d772-catalog-content\") pod \"43c37957-2e0c-44bb-9072-b5bcb256d772\" (UID: \"43c37957-2e0c-44bb-9072-b5bcb256d772\") " Oct 07 15:13:01 crc kubenswrapper[4959]: I1007 15:13:01.296942 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvfg7\" (UniqueName: \"kubernetes.io/projected/43c37957-2e0c-44bb-9072-b5bcb256d772-kube-api-access-xvfg7\") pod \"43c37957-2e0c-44bb-9072-b5bcb256d772\" (UID: \"43c37957-2e0c-44bb-9072-b5bcb256d772\") " Oct 07 15:13:01 crc kubenswrapper[4959]: I1007 15:13:01.298669 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43c37957-2e0c-44bb-9072-b5bcb256d772-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 15:13:01 crc kubenswrapper[4959]: I1007 15:13:01.319873 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43c37957-2e0c-44bb-9072-b5bcb256d772-kube-api-access-xvfg7" (OuterVolumeSpecName: "kube-api-access-xvfg7") pod "43c37957-2e0c-44bb-9072-b5bcb256d772" (UID: "43c37957-2e0c-44bb-9072-b5bcb256d772"). InnerVolumeSpecName "kube-api-access-xvfg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 15:13:01 crc kubenswrapper[4959]: I1007 15:13:01.396248 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43c37957-2e0c-44bb-9072-b5bcb256d772-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "43c37957-2e0c-44bb-9072-b5bcb256d772" (UID: "43c37957-2e0c-44bb-9072-b5bcb256d772"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:13:01 crc kubenswrapper[4959]: I1007 15:13:01.400166 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43c37957-2e0c-44bb-9072-b5bcb256d772-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 15:13:01 crc kubenswrapper[4959]: I1007 15:13:01.400197 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvfg7\" (UniqueName: \"kubernetes.io/projected/43c37957-2e0c-44bb-9072-b5bcb256d772-kube-api-access-xvfg7\") on node \"crc\" DevicePath \"\"" Oct 07 15:13:01 crc kubenswrapper[4959]: I1007 15:13:01.485032 4959 generic.go:334] "Generic (PLEG): container finished" podID="43c37957-2e0c-44bb-9072-b5bcb256d772" containerID="d2bc7480abc71b7a9703c7f6ecc0c12387197b5ac794b7244a382b4fef3e29fd" exitCode=0 Oct 07 15:13:01 crc kubenswrapper[4959]: I1007 15:13:01.485132 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-txh4q" event={"ID":"43c37957-2e0c-44bb-9072-b5bcb256d772","Type":"ContainerDied","Data":"d2bc7480abc71b7a9703c7f6ecc0c12387197b5ac794b7244a382b4fef3e29fd"} Oct 07 15:13:01 crc kubenswrapper[4959]: I1007 15:13:01.485182 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-txh4q" event={"ID":"43c37957-2e0c-44bb-9072-b5bcb256d772","Type":"ContainerDied","Data":"c9e53230e046485637f40b705560bebab5c6135be7168ddf416d276f407ecdd5"} Oct 07 15:13:01 crc kubenswrapper[4959]: I1007 15:13:01.485207 4959 scope.go:117] "RemoveContainer" containerID="d2bc7480abc71b7a9703c7f6ecc0c12387197b5ac794b7244a382b4fef3e29fd" Oct 07 15:13:01 crc kubenswrapper[4959]: I1007 15:13:01.485205 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-txh4q" Oct 07 15:13:01 crc kubenswrapper[4959]: I1007 15:13:01.519584 4959 scope.go:117] "RemoveContainer" containerID="13d489449f923aa540a8b4876a9b4f309be07caba8289eaa4471f4e7bc6c13c2" Oct 07 15:13:01 crc kubenswrapper[4959]: I1007 15:13:01.530704 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-txh4q"] Oct 07 15:13:01 crc kubenswrapper[4959]: I1007 15:13:01.552389 4959 scope.go:117] "RemoveContainer" containerID="d37ecb903482296123abb2583c8d4a87dbc9cad64b64a3fbbc3d729348a4b2e5" Oct 07 15:13:01 crc kubenswrapper[4959]: I1007 15:13:01.553196 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-txh4q"] Oct 07 15:13:01 crc kubenswrapper[4959]: I1007 15:13:01.602144 4959 scope.go:117] "RemoveContainer" containerID="d2bc7480abc71b7a9703c7f6ecc0c12387197b5ac794b7244a382b4fef3e29fd" Oct 07 15:13:01 crc kubenswrapper[4959]: E1007 15:13:01.602624 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2bc7480abc71b7a9703c7f6ecc0c12387197b5ac794b7244a382b4fef3e29fd\": container with ID starting with d2bc7480abc71b7a9703c7f6ecc0c12387197b5ac794b7244a382b4fef3e29fd not found: ID does not exist" containerID="d2bc7480abc71b7a9703c7f6ecc0c12387197b5ac794b7244a382b4fef3e29fd" Oct 07 15:13:01 crc kubenswrapper[4959]: I1007 15:13:01.602685 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2bc7480abc71b7a9703c7f6ecc0c12387197b5ac794b7244a382b4fef3e29fd"} err="failed to get container status \"d2bc7480abc71b7a9703c7f6ecc0c12387197b5ac794b7244a382b4fef3e29fd\": rpc error: code = NotFound desc = could not find container \"d2bc7480abc71b7a9703c7f6ecc0c12387197b5ac794b7244a382b4fef3e29fd\": container with ID starting with d2bc7480abc71b7a9703c7f6ecc0c12387197b5ac794b7244a382b4fef3e29fd not found: ID does not exist" Oct 07 15:13:01 crc kubenswrapper[4959]: I1007 15:13:01.602731 4959 scope.go:117] "RemoveContainer" containerID="13d489449f923aa540a8b4876a9b4f309be07caba8289eaa4471f4e7bc6c13c2" Oct 07 15:13:01 crc kubenswrapper[4959]: E1007 15:13:01.603075 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13d489449f923aa540a8b4876a9b4f309be07caba8289eaa4471f4e7bc6c13c2\": container with ID starting with 13d489449f923aa540a8b4876a9b4f309be07caba8289eaa4471f4e7bc6c13c2 not found: ID does not exist" containerID="13d489449f923aa540a8b4876a9b4f309be07caba8289eaa4471f4e7bc6c13c2" Oct 07 15:13:01 crc kubenswrapper[4959]: I1007 15:13:01.603137 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13d489449f923aa540a8b4876a9b4f309be07caba8289eaa4471f4e7bc6c13c2"} err="failed to get container status \"13d489449f923aa540a8b4876a9b4f309be07caba8289eaa4471f4e7bc6c13c2\": rpc error: code = NotFound desc = could not find container \"13d489449f923aa540a8b4876a9b4f309be07caba8289eaa4471f4e7bc6c13c2\": container with ID starting with 13d489449f923aa540a8b4876a9b4f309be07caba8289eaa4471f4e7bc6c13c2 not found: ID does not exist" Oct 07 15:13:01 crc kubenswrapper[4959]: I1007 15:13:01.603169 4959 scope.go:117] "RemoveContainer" containerID="d37ecb903482296123abb2583c8d4a87dbc9cad64b64a3fbbc3d729348a4b2e5" Oct 07 15:13:01 crc kubenswrapper[4959]: E1007 15:13:01.603698 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d37ecb903482296123abb2583c8d4a87dbc9cad64b64a3fbbc3d729348a4b2e5\": container with ID starting with d37ecb903482296123abb2583c8d4a87dbc9cad64b64a3fbbc3d729348a4b2e5 not found: ID does not exist" containerID="d37ecb903482296123abb2583c8d4a87dbc9cad64b64a3fbbc3d729348a4b2e5" Oct 07 15:13:01 crc kubenswrapper[4959]: I1007 15:13:01.603729 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d37ecb903482296123abb2583c8d4a87dbc9cad64b64a3fbbc3d729348a4b2e5"} err="failed to get container status \"d37ecb903482296123abb2583c8d4a87dbc9cad64b64a3fbbc3d729348a4b2e5\": rpc error: code = NotFound desc = could not find container \"d37ecb903482296123abb2583c8d4a87dbc9cad64b64a3fbbc3d729348a4b2e5\": container with ID starting with d37ecb903482296123abb2583c8d4a87dbc9cad64b64a3fbbc3d729348a4b2e5 not found: ID does not exist" Oct 07 15:13:01 crc kubenswrapper[4959]: I1007 15:13:01.654203 4959 scope.go:117] "RemoveContainer" containerID="e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074" Oct 07 15:13:01 crc kubenswrapper[4959]: E1007 15:13:01.654723 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:13:02 crc kubenswrapper[4959]: I1007 15:13:02.670798 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43c37957-2e0c-44bb-9072-b5bcb256d772" path="/var/lib/kubelet/pods/43c37957-2e0c-44bb-9072-b5bcb256d772/volumes" Oct 07 15:13:15 crc kubenswrapper[4959]: I1007 15:13:15.653739 4959 scope.go:117] "RemoveContainer" containerID="e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074" Oct 07 15:13:15 crc kubenswrapper[4959]: E1007 15:13:15.654991 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:13:26 crc kubenswrapper[4959]: I1007 15:13:26.653941 4959 scope.go:117] "RemoveContainer" containerID="e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074" Oct 07 15:13:26 crc kubenswrapper[4959]: E1007 15:13:26.656240 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:13:41 crc kubenswrapper[4959]: I1007 15:13:41.653693 4959 scope.go:117] "RemoveContainer" containerID="e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074" Oct 07 15:13:41 crc kubenswrapper[4959]: E1007 15:13:41.654552 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:13:55 crc kubenswrapper[4959]: I1007 15:13:55.654025 4959 scope.go:117] "RemoveContainer" containerID="e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074" Oct 07 15:13:55 crc kubenswrapper[4959]: E1007 15:13:55.654847 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:14:09 crc kubenswrapper[4959]: I1007 15:14:09.654117 4959 scope.go:117] "RemoveContainer" containerID="e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074" Oct 07 15:14:09 crc kubenswrapper[4959]: E1007 15:14:09.654879 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:14:22 crc kubenswrapper[4959]: I1007 15:14:22.654549 4959 scope.go:117] "RemoveContainer" containerID="e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074" Oct 07 15:14:23 crc kubenswrapper[4959]: I1007 15:14:23.312048 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerStarted","Data":"2c46cfa18edd8dc1249072876738f373407f694f7fb80d8b963bc7e73086ff89"} Oct 07 15:15:00 crc kubenswrapper[4959]: I1007 15:15:00.147352 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330835-7t6n7"] Oct 07 15:15:00 crc kubenswrapper[4959]: E1007 15:15:00.151478 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43c37957-2e0c-44bb-9072-b5bcb256d772" containerName="extract-utilities" Oct 07 15:15:00 crc kubenswrapper[4959]: I1007 15:15:00.151677 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="43c37957-2e0c-44bb-9072-b5bcb256d772" containerName="extract-utilities" Oct 07 15:15:00 crc kubenswrapper[4959]: E1007 15:15:00.151764 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43c37957-2e0c-44bb-9072-b5bcb256d772" containerName="registry-server" Oct 07 15:15:00 crc kubenswrapper[4959]: I1007 15:15:00.151840 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="43c37957-2e0c-44bb-9072-b5bcb256d772" containerName="registry-server" Oct 07 15:15:00 crc kubenswrapper[4959]: E1007 15:15:00.151935 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43c37957-2e0c-44bb-9072-b5bcb256d772" containerName="extract-content" Oct 07 15:15:00 crc kubenswrapper[4959]: I1007 15:15:00.152013 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="43c37957-2e0c-44bb-9072-b5bcb256d772" containerName="extract-content" Oct 07 15:15:00 crc kubenswrapper[4959]: I1007 15:15:00.152453 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="43c37957-2e0c-44bb-9072-b5bcb256d772" containerName="registry-server" Oct 07 15:15:00 crc kubenswrapper[4959]: I1007 15:15:00.153491 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330835-7t6n7" Oct 07 15:15:00 crc kubenswrapper[4959]: I1007 15:15:00.156117 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 07 15:15:00 crc kubenswrapper[4959]: I1007 15:15:00.156161 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 07 15:15:00 crc kubenswrapper[4959]: I1007 15:15:00.158701 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330835-7t6n7"] Oct 07 15:15:00 crc kubenswrapper[4959]: I1007 15:15:00.192881 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztlt7\" (UniqueName: \"kubernetes.io/projected/f2dd2f4a-4231-44a9-87fa-99f90bef828d-kube-api-access-ztlt7\") pod \"collect-profiles-29330835-7t6n7\" (UID: \"f2dd2f4a-4231-44a9-87fa-99f90bef828d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330835-7t6n7" Oct 07 15:15:00 crc kubenswrapper[4959]: I1007 15:15:00.192979 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2dd2f4a-4231-44a9-87fa-99f90bef828d-config-volume\") pod \"collect-profiles-29330835-7t6n7\" (UID: \"f2dd2f4a-4231-44a9-87fa-99f90bef828d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330835-7t6n7" Oct 07 15:15:00 crc kubenswrapper[4959]: I1007 15:15:00.193229 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2dd2f4a-4231-44a9-87fa-99f90bef828d-secret-volume\") pod \"collect-profiles-29330835-7t6n7\" (UID: \"f2dd2f4a-4231-44a9-87fa-99f90bef828d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330835-7t6n7" Oct 07 15:15:00 crc kubenswrapper[4959]: I1007 15:15:00.295229 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztlt7\" (UniqueName: \"kubernetes.io/projected/f2dd2f4a-4231-44a9-87fa-99f90bef828d-kube-api-access-ztlt7\") pod \"collect-profiles-29330835-7t6n7\" (UID: \"f2dd2f4a-4231-44a9-87fa-99f90bef828d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330835-7t6n7" Oct 07 15:15:00 crc kubenswrapper[4959]: I1007 15:15:00.295386 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2dd2f4a-4231-44a9-87fa-99f90bef828d-config-volume\") pod \"collect-profiles-29330835-7t6n7\" (UID: \"f2dd2f4a-4231-44a9-87fa-99f90bef828d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330835-7t6n7" Oct 07 15:15:00 crc kubenswrapper[4959]: I1007 15:15:00.295481 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2dd2f4a-4231-44a9-87fa-99f90bef828d-secret-volume\") pod \"collect-profiles-29330835-7t6n7\" (UID: \"f2dd2f4a-4231-44a9-87fa-99f90bef828d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330835-7t6n7" Oct 07 15:15:00 crc kubenswrapper[4959]: I1007 15:15:00.296308 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2dd2f4a-4231-44a9-87fa-99f90bef828d-config-volume\") pod \"collect-profiles-29330835-7t6n7\" (UID: \"f2dd2f4a-4231-44a9-87fa-99f90bef828d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330835-7t6n7" Oct 07 15:15:00 crc kubenswrapper[4959]: I1007 15:15:00.301121 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2dd2f4a-4231-44a9-87fa-99f90bef828d-secret-volume\") pod \"collect-profiles-29330835-7t6n7\" (UID: \"f2dd2f4a-4231-44a9-87fa-99f90bef828d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330835-7t6n7" Oct 07 15:15:00 crc kubenswrapper[4959]: I1007 15:15:00.316260 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztlt7\" (UniqueName: \"kubernetes.io/projected/f2dd2f4a-4231-44a9-87fa-99f90bef828d-kube-api-access-ztlt7\") pod \"collect-profiles-29330835-7t6n7\" (UID: \"f2dd2f4a-4231-44a9-87fa-99f90bef828d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330835-7t6n7" Oct 07 15:15:00 crc kubenswrapper[4959]: I1007 15:15:00.489023 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330835-7t6n7" Oct 07 15:15:00 crc kubenswrapper[4959]: I1007 15:15:00.966186 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330835-7t6n7"] Oct 07 15:15:01 crc kubenswrapper[4959]: I1007 15:15:01.685950 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330835-7t6n7" event={"ID":"f2dd2f4a-4231-44a9-87fa-99f90bef828d","Type":"ContainerStarted","Data":"3a7717a2af2228ee0fa0039615430c938f39969edc82304ad301dff3216fa9d6"} Oct 07 15:15:01 crc kubenswrapper[4959]: I1007 15:15:01.685997 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330835-7t6n7" event={"ID":"f2dd2f4a-4231-44a9-87fa-99f90bef828d","Type":"ContainerStarted","Data":"107cb8f701a60609cbcdabd70f4bf427c0063c0967e0e21378ca8287b8b7c112"} Oct 07 15:15:01 crc kubenswrapper[4959]: I1007 15:15:01.714014 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29330835-7t6n7" podStartSLOduration=1.713990466 podStartE2EDuration="1.713990466s" podCreationTimestamp="2025-10-07 15:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 15:15:01.703690623 +0000 UTC m=+5383.787094948" watchObservedRunningTime="2025-10-07 15:15:01.713990466 +0000 UTC m=+5383.797394791" Oct 07 15:15:02 crc kubenswrapper[4959]: I1007 15:15:02.696214 4959 generic.go:334] "Generic (PLEG): container finished" podID="f2dd2f4a-4231-44a9-87fa-99f90bef828d" containerID="3a7717a2af2228ee0fa0039615430c938f39969edc82304ad301dff3216fa9d6" exitCode=0 Oct 07 15:15:02 crc kubenswrapper[4959]: I1007 15:15:02.696324 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330835-7t6n7" event={"ID":"f2dd2f4a-4231-44a9-87fa-99f90bef828d","Type":"ContainerDied","Data":"3a7717a2af2228ee0fa0039615430c938f39969edc82304ad301dff3216fa9d6"} Oct 07 15:15:04 crc kubenswrapper[4959]: I1007 15:15:04.169466 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330835-7t6n7" Oct 07 15:15:04 crc kubenswrapper[4959]: I1007 15:15:04.181510 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztlt7\" (UniqueName: \"kubernetes.io/projected/f2dd2f4a-4231-44a9-87fa-99f90bef828d-kube-api-access-ztlt7\") pod \"f2dd2f4a-4231-44a9-87fa-99f90bef828d\" (UID: \"f2dd2f4a-4231-44a9-87fa-99f90bef828d\") " Oct 07 15:15:04 crc kubenswrapper[4959]: I1007 15:15:04.181787 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2dd2f4a-4231-44a9-87fa-99f90bef828d-secret-volume\") pod \"f2dd2f4a-4231-44a9-87fa-99f90bef828d\" (UID: \"f2dd2f4a-4231-44a9-87fa-99f90bef828d\") " Oct 07 15:15:04 crc kubenswrapper[4959]: I1007 15:15:04.181889 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2dd2f4a-4231-44a9-87fa-99f90bef828d-config-volume\") pod \"f2dd2f4a-4231-44a9-87fa-99f90bef828d\" (UID: \"f2dd2f4a-4231-44a9-87fa-99f90bef828d\") " Oct 07 15:15:04 crc kubenswrapper[4959]: I1007 15:15:04.182551 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2dd2f4a-4231-44a9-87fa-99f90bef828d-config-volume" (OuterVolumeSpecName: "config-volume") pod "f2dd2f4a-4231-44a9-87fa-99f90bef828d" (UID: "f2dd2f4a-4231-44a9-87fa-99f90bef828d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 15:15:04 crc kubenswrapper[4959]: I1007 15:15:04.183015 4959 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2dd2f4a-4231-44a9-87fa-99f90bef828d-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 15:15:04 crc kubenswrapper[4959]: I1007 15:15:04.189681 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2dd2f4a-4231-44a9-87fa-99f90bef828d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f2dd2f4a-4231-44a9-87fa-99f90bef828d" (UID: "f2dd2f4a-4231-44a9-87fa-99f90bef828d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:15:04 crc kubenswrapper[4959]: I1007 15:15:04.189885 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2dd2f4a-4231-44a9-87fa-99f90bef828d-kube-api-access-ztlt7" (OuterVolumeSpecName: "kube-api-access-ztlt7") pod "f2dd2f4a-4231-44a9-87fa-99f90bef828d" (UID: "f2dd2f4a-4231-44a9-87fa-99f90bef828d"). InnerVolumeSpecName "kube-api-access-ztlt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 15:15:04 crc kubenswrapper[4959]: I1007 15:15:04.283678 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztlt7\" (UniqueName: \"kubernetes.io/projected/f2dd2f4a-4231-44a9-87fa-99f90bef828d-kube-api-access-ztlt7\") on node \"crc\" DevicePath \"\"" Oct 07 15:15:04 crc kubenswrapper[4959]: I1007 15:15:04.283988 4959 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2dd2f4a-4231-44a9-87fa-99f90bef828d-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 07 15:15:04 crc kubenswrapper[4959]: I1007 15:15:04.717157 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330835-7t6n7" event={"ID":"f2dd2f4a-4231-44a9-87fa-99f90bef828d","Type":"ContainerDied","Data":"107cb8f701a60609cbcdabd70f4bf427c0063c0967e0e21378ca8287b8b7c112"} Oct 07 15:15:04 crc kubenswrapper[4959]: I1007 15:15:04.717204 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="107cb8f701a60609cbcdabd70f4bf427c0063c0967e0e21378ca8287b8b7c112" Oct 07 15:15:04 crc kubenswrapper[4959]: I1007 15:15:04.717215 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330835-7t6n7" Oct 07 15:15:04 crc kubenswrapper[4959]: I1007 15:15:04.783881 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330790-nrtvc"] Oct 07 15:15:04 crc kubenswrapper[4959]: I1007 15:15:04.796158 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330790-nrtvc"] Oct 07 15:15:06 crc kubenswrapper[4959]: I1007 15:15:06.673463 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab" path="/var/lib/kubelet/pods/2bfd55bc-a9f8-4726-b5d9-588ce4ff8fab/volumes" Oct 07 15:15:32 crc kubenswrapper[4959]: I1007 15:15:32.327441 4959 scope.go:117] "RemoveContainer" containerID="766f0cd77ccad16e5276ddfc64d80202ceae10e653ab4a0fbcdd693e1a845aa1" Oct 07 15:15:33 crc kubenswrapper[4959]: I1007 15:15:33.701224 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tdctl"] Oct 07 15:15:33 crc kubenswrapper[4959]: E1007 15:15:33.701886 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2dd2f4a-4231-44a9-87fa-99f90bef828d" containerName="collect-profiles" Oct 07 15:15:33 crc kubenswrapper[4959]: I1007 15:15:33.701903 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2dd2f4a-4231-44a9-87fa-99f90bef828d" containerName="collect-profiles" Oct 07 15:15:33 crc kubenswrapper[4959]: I1007 15:15:33.702112 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2dd2f4a-4231-44a9-87fa-99f90bef828d" containerName="collect-profiles" Oct 07 15:15:33 crc kubenswrapper[4959]: I1007 15:15:33.708972 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tdctl" Oct 07 15:15:33 crc kubenswrapper[4959]: I1007 15:15:33.735329 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tdctl"] Oct 07 15:15:33 crc kubenswrapper[4959]: I1007 15:15:33.867587 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ljvf\" (UniqueName: \"kubernetes.io/projected/35d3811a-4b30-4a77-ae2a-f43113b1ebfc-kube-api-access-6ljvf\") pod \"redhat-operators-tdctl\" (UID: \"35d3811a-4b30-4a77-ae2a-f43113b1ebfc\") " pod="openshift-marketplace/redhat-operators-tdctl" Oct 07 15:15:33 crc kubenswrapper[4959]: I1007 15:15:33.867851 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35d3811a-4b30-4a77-ae2a-f43113b1ebfc-catalog-content\") pod \"redhat-operators-tdctl\" (UID: \"35d3811a-4b30-4a77-ae2a-f43113b1ebfc\") " pod="openshift-marketplace/redhat-operators-tdctl" Oct 07 15:15:33 crc kubenswrapper[4959]: I1007 15:15:33.868022 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35d3811a-4b30-4a77-ae2a-f43113b1ebfc-utilities\") pod \"redhat-operators-tdctl\" (UID: \"35d3811a-4b30-4a77-ae2a-f43113b1ebfc\") " pod="openshift-marketplace/redhat-operators-tdctl" Oct 07 15:15:33 crc kubenswrapper[4959]: I1007 15:15:33.970049 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35d3811a-4b30-4a77-ae2a-f43113b1ebfc-utilities\") pod \"redhat-operators-tdctl\" (UID: \"35d3811a-4b30-4a77-ae2a-f43113b1ebfc\") " pod="openshift-marketplace/redhat-operators-tdctl" Oct 07 15:15:33 crc kubenswrapper[4959]: I1007 15:15:33.970238 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ljvf\" (UniqueName: \"kubernetes.io/projected/35d3811a-4b30-4a77-ae2a-f43113b1ebfc-kube-api-access-6ljvf\") pod \"redhat-operators-tdctl\" (UID: \"35d3811a-4b30-4a77-ae2a-f43113b1ebfc\") " pod="openshift-marketplace/redhat-operators-tdctl" Oct 07 15:15:33 crc kubenswrapper[4959]: I1007 15:15:33.970317 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35d3811a-4b30-4a77-ae2a-f43113b1ebfc-catalog-content\") pod \"redhat-operators-tdctl\" (UID: \"35d3811a-4b30-4a77-ae2a-f43113b1ebfc\") " pod="openshift-marketplace/redhat-operators-tdctl" Oct 07 15:15:33 crc kubenswrapper[4959]: I1007 15:15:33.970847 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35d3811a-4b30-4a77-ae2a-f43113b1ebfc-catalog-content\") pod \"redhat-operators-tdctl\" (UID: \"35d3811a-4b30-4a77-ae2a-f43113b1ebfc\") " pod="openshift-marketplace/redhat-operators-tdctl" Oct 07 15:15:33 crc kubenswrapper[4959]: I1007 15:15:33.970949 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35d3811a-4b30-4a77-ae2a-f43113b1ebfc-utilities\") pod \"redhat-operators-tdctl\" (UID: \"35d3811a-4b30-4a77-ae2a-f43113b1ebfc\") " pod="openshift-marketplace/redhat-operators-tdctl" Oct 07 15:15:33 crc kubenswrapper[4959]: I1007 15:15:33.995239 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ljvf\" (UniqueName: \"kubernetes.io/projected/35d3811a-4b30-4a77-ae2a-f43113b1ebfc-kube-api-access-6ljvf\") pod \"redhat-operators-tdctl\" (UID: \"35d3811a-4b30-4a77-ae2a-f43113b1ebfc\") " pod="openshift-marketplace/redhat-operators-tdctl" Oct 07 15:15:34 crc kubenswrapper[4959]: I1007 15:15:34.036157 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tdctl" Oct 07 15:15:34 crc kubenswrapper[4959]: I1007 15:15:34.588815 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tdctl"] Oct 07 15:15:34 crc kubenswrapper[4959]: I1007 15:15:34.987766 4959 generic.go:334] "Generic (PLEG): container finished" podID="35d3811a-4b30-4a77-ae2a-f43113b1ebfc" containerID="d4c0dfa6c37088a8cb53a5c7b0e6488aa97192dce079f6562a36d56affc6d0b8" exitCode=0 Oct 07 15:15:34 crc kubenswrapper[4959]: I1007 15:15:34.987885 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tdctl" event={"ID":"35d3811a-4b30-4a77-ae2a-f43113b1ebfc","Type":"ContainerDied","Data":"d4c0dfa6c37088a8cb53a5c7b0e6488aa97192dce079f6562a36d56affc6d0b8"} Oct 07 15:15:34 crc kubenswrapper[4959]: I1007 15:15:34.988218 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tdctl" event={"ID":"35d3811a-4b30-4a77-ae2a-f43113b1ebfc","Type":"ContainerStarted","Data":"e386575d39668e97950bcf60c073eb437e3c0988b81b4963fe20dee9aced7677"} Oct 07 15:15:38 crc kubenswrapper[4959]: I1007 15:15:38.017240 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tdctl" event={"ID":"35d3811a-4b30-4a77-ae2a-f43113b1ebfc","Type":"ContainerStarted","Data":"c35a30b993166b484c773678ef0ec73cb8cac2a17b4c0785adbd5ebc33000c9d"} Oct 07 15:15:45 crc kubenswrapper[4959]: I1007 15:15:45.081981 4959 generic.go:334] "Generic (PLEG): container finished" podID="35d3811a-4b30-4a77-ae2a-f43113b1ebfc" containerID="c35a30b993166b484c773678ef0ec73cb8cac2a17b4c0785adbd5ebc33000c9d" exitCode=0 Oct 07 15:15:45 crc kubenswrapper[4959]: I1007 15:15:45.082065 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tdctl" event={"ID":"35d3811a-4b30-4a77-ae2a-f43113b1ebfc","Type":"ContainerDied","Data":"c35a30b993166b484c773678ef0ec73cb8cac2a17b4c0785adbd5ebc33000c9d"} Oct 07 15:15:46 crc kubenswrapper[4959]: I1007 15:15:46.096180 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tdctl" event={"ID":"35d3811a-4b30-4a77-ae2a-f43113b1ebfc","Type":"ContainerStarted","Data":"d7a8198f7611fff8da4017dafcd54e705cc048801d7532e01b77b94afd9a774d"} Oct 07 15:15:46 crc kubenswrapper[4959]: I1007 15:15:46.124455 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tdctl" podStartSLOduration=2.497539279 podStartE2EDuration="13.124427598s" podCreationTimestamp="2025-10-07 15:15:33 +0000 UTC" firstStartedPulling="2025-10-07 15:15:34.98988246 +0000 UTC m=+5417.073286785" lastFinishedPulling="2025-10-07 15:15:45.616770779 +0000 UTC m=+5427.700175104" observedRunningTime="2025-10-07 15:15:46.118431154 +0000 UTC m=+5428.201835489" watchObservedRunningTime="2025-10-07 15:15:46.124427598 +0000 UTC m=+5428.207831913" Oct 07 15:15:54 crc kubenswrapper[4959]: I1007 15:15:54.037223 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tdctl" Oct 07 15:15:54 crc kubenswrapper[4959]: I1007 15:15:54.038909 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tdctl" Oct 07 15:15:54 crc kubenswrapper[4959]: I1007 15:15:54.085073 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tdctl" Oct 07 15:15:54 crc kubenswrapper[4959]: I1007 15:15:54.259771 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tdctl" Oct 07 15:15:54 crc kubenswrapper[4959]: I1007 15:15:54.362136 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tdctl"] Oct 07 15:15:56 crc kubenswrapper[4959]: I1007 15:15:56.210919 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tdctl" podUID="35d3811a-4b30-4a77-ae2a-f43113b1ebfc" containerName="registry-server" containerID="cri-o://d7a8198f7611fff8da4017dafcd54e705cc048801d7532e01b77b94afd9a774d" gracePeriod=2 Oct 07 15:15:56 crc kubenswrapper[4959]: I1007 15:15:56.744823 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tdctl" Oct 07 15:15:56 crc kubenswrapper[4959]: I1007 15:15:56.815936 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35d3811a-4b30-4a77-ae2a-f43113b1ebfc-catalog-content\") pod \"35d3811a-4b30-4a77-ae2a-f43113b1ebfc\" (UID: \"35d3811a-4b30-4a77-ae2a-f43113b1ebfc\") " Oct 07 15:15:56 crc kubenswrapper[4959]: I1007 15:15:56.818372 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ljvf\" (UniqueName: \"kubernetes.io/projected/35d3811a-4b30-4a77-ae2a-f43113b1ebfc-kube-api-access-6ljvf\") pod \"35d3811a-4b30-4a77-ae2a-f43113b1ebfc\" (UID: \"35d3811a-4b30-4a77-ae2a-f43113b1ebfc\") " Oct 07 15:15:56 crc kubenswrapper[4959]: I1007 15:15:56.818695 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35d3811a-4b30-4a77-ae2a-f43113b1ebfc-utilities\") pod \"35d3811a-4b30-4a77-ae2a-f43113b1ebfc\" (UID: \"35d3811a-4b30-4a77-ae2a-f43113b1ebfc\") " Oct 07 15:15:56 crc kubenswrapper[4959]: I1007 15:15:56.819845 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35d3811a-4b30-4a77-ae2a-f43113b1ebfc-utilities" (OuterVolumeSpecName: "utilities") pod "35d3811a-4b30-4a77-ae2a-f43113b1ebfc" (UID: "35d3811a-4b30-4a77-ae2a-f43113b1ebfc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:15:56 crc kubenswrapper[4959]: I1007 15:15:56.838889 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35d3811a-4b30-4a77-ae2a-f43113b1ebfc-kube-api-access-6ljvf" (OuterVolumeSpecName: "kube-api-access-6ljvf") pod "35d3811a-4b30-4a77-ae2a-f43113b1ebfc" (UID: "35d3811a-4b30-4a77-ae2a-f43113b1ebfc"). InnerVolumeSpecName "kube-api-access-6ljvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 15:15:56 crc kubenswrapper[4959]: I1007 15:15:56.922604 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35d3811a-4b30-4a77-ae2a-f43113b1ebfc-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 15:15:56 crc kubenswrapper[4959]: I1007 15:15:56.922662 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ljvf\" (UniqueName: \"kubernetes.io/projected/35d3811a-4b30-4a77-ae2a-f43113b1ebfc-kube-api-access-6ljvf\") on node \"crc\" DevicePath \"\"" Oct 07 15:15:56 crc kubenswrapper[4959]: I1007 15:15:56.926042 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35d3811a-4b30-4a77-ae2a-f43113b1ebfc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "35d3811a-4b30-4a77-ae2a-f43113b1ebfc" (UID: "35d3811a-4b30-4a77-ae2a-f43113b1ebfc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:15:57 crc kubenswrapper[4959]: I1007 15:15:57.025095 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35d3811a-4b30-4a77-ae2a-f43113b1ebfc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 15:15:57 crc kubenswrapper[4959]: I1007 15:15:57.220802 4959 generic.go:334] "Generic (PLEG): container finished" podID="35d3811a-4b30-4a77-ae2a-f43113b1ebfc" containerID="d7a8198f7611fff8da4017dafcd54e705cc048801d7532e01b77b94afd9a774d" exitCode=0 Oct 07 15:15:57 crc kubenswrapper[4959]: I1007 15:15:57.220850 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tdctl" event={"ID":"35d3811a-4b30-4a77-ae2a-f43113b1ebfc","Type":"ContainerDied","Data":"d7a8198f7611fff8da4017dafcd54e705cc048801d7532e01b77b94afd9a774d"} Oct 07 15:15:57 crc kubenswrapper[4959]: I1007 15:15:57.220882 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tdctl" event={"ID":"35d3811a-4b30-4a77-ae2a-f43113b1ebfc","Type":"ContainerDied","Data":"e386575d39668e97950bcf60c073eb437e3c0988b81b4963fe20dee9aced7677"} Oct 07 15:15:57 crc kubenswrapper[4959]: I1007 15:15:57.220900 4959 scope.go:117] "RemoveContainer" containerID="d7a8198f7611fff8da4017dafcd54e705cc048801d7532e01b77b94afd9a774d" Oct 07 15:15:57 crc kubenswrapper[4959]: I1007 15:15:57.221121 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tdctl" Oct 07 15:15:57 crc kubenswrapper[4959]: I1007 15:15:57.254196 4959 scope.go:117] "RemoveContainer" containerID="c35a30b993166b484c773678ef0ec73cb8cac2a17b4c0785adbd5ebc33000c9d" Oct 07 15:15:57 crc kubenswrapper[4959]: I1007 15:15:57.279314 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tdctl"] Oct 07 15:15:57 crc kubenswrapper[4959]: I1007 15:15:57.289352 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tdctl"] Oct 07 15:15:57 crc kubenswrapper[4959]: I1007 15:15:57.291400 4959 scope.go:117] "RemoveContainer" containerID="d4c0dfa6c37088a8cb53a5c7b0e6488aa97192dce079f6562a36d56affc6d0b8" Oct 07 15:15:57 crc kubenswrapper[4959]: I1007 15:15:57.344225 4959 scope.go:117] "RemoveContainer" containerID="d7a8198f7611fff8da4017dafcd54e705cc048801d7532e01b77b94afd9a774d" Oct 07 15:15:57 crc kubenswrapper[4959]: E1007 15:15:57.344782 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7a8198f7611fff8da4017dafcd54e705cc048801d7532e01b77b94afd9a774d\": container with ID starting with d7a8198f7611fff8da4017dafcd54e705cc048801d7532e01b77b94afd9a774d not found: ID does not exist" containerID="d7a8198f7611fff8da4017dafcd54e705cc048801d7532e01b77b94afd9a774d" Oct 07 15:15:57 crc kubenswrapper[4959]: I1007 15:15:57.344832 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7a8198f7611fff8da4017dafcd54e705cc048801d7532e01b77b94afd9a774d"} err="failed to get container status \"d7a8198f7611fff8da4017dafcd54e705cc048801d7532e01b77b94afd9a774d\": rpc error: code = NotFound desc = could not find container \"d7a8198f7611fff8da4017dafcd54e705cc048801d7532e01b77b94afd9a774d\": container with ID starting with d7a8198f7611fff8da4017dafcd54e705cc048801d7532e01b77b94afd9a774d not found: ID does not exist" Oct 07 15:15:57 crc kubenswrapper[4959]: I1007 15:15:57.344864 4959 scope.go:117] "RemoveContainer" containerID="c35a30b993166b484c773678ef0ec73cb8cac2a17b4c0785adbd5ebc33000c9d" Oct 07 15:15:57 crc kubenswrapper[4959]: E1007 15:15:57.345398 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c35a30b993166b484c773678ef0ec73cb8cac2a17b4c0785adbd5ebc33000c9d\": container with ID starting with c35a30b993166b484c773678ef0ec73cb8cac2a17b4c0785adbd5ebc33000c9d not found: ID does not exist" containerID="c35a30b993166b484c773678ef0ec73cb8cac2a17b4c0785adbd5ebc33000c9d" Oct 07 15:15:57 crc kubenswrapper[4959]: I1007 15:15:57.345436 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c35a30b993166b484c773678ef0ec73cb8cac2a17b4c0785adbd5ebc33000c9d"} err="failed to get container status \"c35a30b993166b484c773678ef0ec73cb8cac2a17b4c0785adbd5ebc33000c9d\": rpc error: code = NotFound desc = could not find container \"c35a30b993166b484c773678ef0ec73cb8cac2a17b4c0785adbd5ebc33000c9d\": container with ID starting with c35a30b993166b484c773678ef0ec73cb8cac2a17b4c0785adbd5ebc33000c9d not found: ID does not exist" Oct 07 15:15:57 crc kubenswrapper[4959]: I1007 15:15:57.345468 4959 scope.go:117] "RemoveContainer" containerID="d4c0dfa6c37088a8cb53a5c7b0e6488aa97192dce079f6562a36d56affc6d0b8" Oct 07 15:15:57 crc kubenswrapper[4959]: E1007 15:15:57.345984 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4c0dfa6c37088a8cb53a5c7b0e6488aa97192dce079f6562a36d56affc6d0b8\": container with ID starting with d4c0dfa6c37088a8cb53a5c7b0e6488aa97192dce079f6562a36d56affc6d0b8 not found: ID does not exist" containerID="d4c0dfa6c37088a8cb53a5c7b0e6488aa97192dce079f6562a36d56affc6d0b8" Oct 07 15:15:57 crc kubenswrapper[4959]: I1007 15:15:57.346051 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4c0dfa6c37088a8cb53a5c7b0e6488aa97192dce079f6562a36d56affc6d0b8"} err="failed to get container status \"d4c0dfa6c37088a8cb53a5c7b0e6488aa97192dce079f6562a36d56affc6d0b8\": rpc error: code = NotFound desc = could not find container \"d4c0dfa6c37088a8cb53a5c7b0e6488aa97192dce079f6562a36d56affc6d0b8\": container with ID starting with d4c0dfa6c37088a8cb53a5c7b0e6488aa97192dce079f6562a36d56affc6d0b8 not found: ID does not exist" Oct 07 15:15:58 crc kubenswrapper[4959]: I1007 15:15:58.664916 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35d3811a-4b30-4a77-ae2a-f43113b1ebfc" path="/var/lib/kubelet/pods/35d3811a-4b30-4a77-ae2a-f43113b1ebfc/volumes" Oct 07 15:16:50 crc kubenswrapper[4959]: I1007 15:16:50.630585 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 15:16:50 crc kubenswrapper[4959]: I1007 15:16:50.633346 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 15:17:20 crc kubenswrapper[4959]: I1007 15:17:20.630867 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 15:17:20 crc kubenswrapper[4959]: I1007 15:17:20.631987 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 15:17:28 crc kubenswrapper[4959]: I1007 15:17:28.159606 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f2vz2"] Oct 07 15:17:28 crc kubenswrapper[4959]: E1007 15:17:28.161587 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35d3811a-4b30-4a77-ae2a-f43113b1ebfc" containerName="registry-server" Oct 07 15:17:28 crc kubenswrapper[4959]: I1007 15:17:28.161680 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="35d3811a-4b30-4a77-ae2a-f43113b1ebfc" containerName="registry-server" Oct 07 15:17:28 crc kubenswrapper[4959]: E1007 15:17:28.161752 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35d3811a-4b30-4a77-ae2a-f43113b1ebfc" containerName="extract-content" Oct 07 15:17:28 crc kubenswrapper[4959]: I1007 15:17:28.161808 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="35d3811a-4b30-4a77-ae2a-f43113b1ebfc" containerName="extract-content" Oct 07 15:17:28 crc kubenswrapper[4959]: E1007 15:17:28.161881 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35d3811a-4b30-4a77-ae2a-f43113b1ebfc" containerName="extract-utilities" Oct 07 15:17:28 crc kubenswrapper[4959]: I1007 15:17:28.161938 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="35d3811a-4b30-4a77-ae2a-f43113b1ebfc" containerName="extract-utilities" Oct 07 15:17:28 crc kubenswrapper[4959]: I1007 15:17:28.162260 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="35d3811a-4b30-4a77-ae2a-f43113b1ebfc" containerName="registry-server" Oct 07 15:17:28 crc kubenswrapper[4959]: I1007 15:17:28.163933 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f2vz2" Oct 07 15:17:28 crc kubenswrapper[4959]: I1007 15:17:28.172314 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f2vz2"] Oct 07 15:17:28 crc kubenswrapper[4959]: I1007 15:17:28.255470 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8dpk\" (UniqueName: \"kubernetes.io/projected/2bf22139-fc42-4993-97f7-0e1db00cc684-kube-api-access-z8dpk\") pod \"community-operators-f2vz2\" (UID: \"2bf22139-fc42-4993-97f7-0e1db00cc684\") " pod="openshift-marketplace/community-operators-f2vz2" Oct 07 15:17:28 crc kubenswrapper[4959]: I1007 15:17:28.255552 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bf22139-fc42-4993-97f7-0e1db00cc684-catalog-content\") pod \"community-operators-f2vz2\" (UID: \"2bf22139-fc42-4993-97f7-0e1db00cc684\") " pod="openshift-marketplace/community-operators-f2vz2" Oct 07 15:17:28 crc kubenswrapper[4959]: I1007 15:17:28.255909 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bf22139-fc42-4993-97f7-0e1db00cc684-utilities\") pod \"community-operators-f2vz2\" (UID: \"2bf22139-fc42-4993-97f7-0e1db00cc684\") " pod="openshift-marketplace/community-operators-f2vz2" Oct 07 15:17:28 crc kubenswrapper[4959]: I1007 15:17:28.358393 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bf22139-fc42-4993-97f7-0e1db00cc684-catalog-content\") pod \"community-operators-f2vz2\" (UID: \"2bf22139-fc42-4993-97f7-0e1db00cc684\") " pod="openshift-marketplace/community-operators-f2vz2" Oct 07 15:17:28 crc kubenswrapper[4959]: I1007 15:17:28.358588 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bf22139-fc42-4993-97f7-0e1db00cc684-utilities\") pod \"community-operators-f2vz2\" (UID: \"2bf22139-fc42-4993-97f7-0e1db00cc684\") " pod="openshift-marketplace/community-operators-f2vz2" Oct 07 15:17:28 crc kubenswrapper[4959]: I1007 15:17:28.358776 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8dpk\" (UniqueName: \"kubernetes.io/projected/2bf22139-fc42-4993-97f7-0e1db00cc684-kube-api-access-z8dpk\") pod \"community-operators-f2vz2\" (UID: \"2bf22139-fc42-4993-97f7-0e1db00cc684\") " pod="openshift-marketplace/community-operators-f2vz2" Oct 07 15:17:28 crc kubenswrapper[4959]: I1007 15:17:28.358984 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bf22139-fc42-4993-97f7-0e1db00cc684-catalog-content\") pod \"community-operators-f2vz2\" (UID: \"2bf22139-fc42-4993-97f7-0e1db00cc684\") " pod="openshift-marketplace/community-operators-f2vz2" Oct 07 15:17:28 crc kubenswrapper[4959]: I1007 15:17:28.359143 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bf22139-fc42-4993-97f7-0e1db00cc684-utilities\") pod \"community-operators-f2vz2\" (UID: \"2bf22139-fc42-4993-97f7-0e1db00cc684\") " pod="openshift-marketplace/community-operators-f2vz2" Oct 07 15:17:28 crc kubenswrapper[4959]: I1007 15:17:28.382310 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8dpk\" (UniqueName: \"kubernetes.io/projected/2bf22139-fc42-4993-97f7-0e1db00cc684-kube-api-access-z8dpk\") pod \"community-operators-f2vz2\" (UID: \"2bf22139-fc42-4993-97f7-0e1db00cc684\") " pod="openshift-marketplace/community-operators-f2vz2" Oct 07 15:17:28 crc kubenswrapper[4959]: I1007 15:17:28.487581 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f2vz2" Oct 07 15:17:29 crc kubenswrapper[4959]: I1007 15:17:29.057547 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f2vz2"] Oct 07 15:17:29 crc kubenswrapper[4959]: I1007 15:17:29.115968 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2vz2" event={"ID":"2bf22139-fc42-4993-97f7-0e1db00cc684","Type":"ContainerStarted","Data":"c76c70a1c837be0093d01bddcd1816828113116d9d4ec92fea4ab28b458b276d"} Oct 07 15:17:30 crc kubenswrapper[4959]: I1007 15:17:30.126495 4959 generic.go:334] "Generic (PLEG): container finished" podID="2bf22139-fc42-4993-97f7-0e1db00cc684" containerID="b3d124a4ccdf6ce3f6af9f744f4663e49a9701341b22289f6ada766b66b3b74a" exitCode=0 Oct 07 15:17:30 crc kubenswrapper[4959]: I1007 15:17:30.126806 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2vz2" event={"ID":"2bf22139-fc42-4993-97f7-0e1db00cc684","Type":"ContainerDied","Data":"b3d124a4ccdf6ce3f6af9f744f4663e49a9701341b22289f6ada766b66b3b74a"} Oct 07 15:17:32 crc kubenswrapper[4959]: I1007 15:17:32.145790 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2vz2" event={"ID":"2bf22139-fc42-4993-97f7-0e1db00cc684","Type":"ContainerStarted","Data":"28abefb82d22403588ad3046e81a173d36a10cca387cc444629e749bd9496c74"} Oct 07 15:17:33 crc kubenswrapper[4959]: I1007 15:17:33.158155 4959 generic.go:334] "Generic (PLEG): container finished" podID="2bf22139-fc42-4993-97f7-0e1db00cc684" containerID="28abefb82d22403588ad3046e81a173d36a10cca387cc444629e749bd9496c74" exitCode=0 Oct 07 15:17:33 crc kubenswrapper[4959]: I1007 15:17:33.158993 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2vz2" event={"ID":"2bf22139-fc42-4993-97f7-0e1db00cc684","Type":"ContainerDied","Data":"28abefb82d22403588ad3046e81a173d36a10cca387cc444629e749bd9496c74"} Oct 07 15:17:34 crc kubenswrapper[4959]: I1007 15:17:34.170264 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2vz2" event={"ID":"2bf22139-fc42-4993-97f7-0e1db00cc684","Type":"ContainerStarted","Data":"90e025308323baa19adb527cb7bc4a706e065a34ea9f2cbb981a7348b9d8cd82"} Oct 07 15:17:34 crc kubenswrapper[4959]: I1007 15:17:34.196779 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f2vz2" podStartSLOduration=2.590080199 podStartE2EDuration="6.196756635s" podCreationTimestamp="2025-10-07 15:17:28 +0000 UTC" firstStartedPulling="2025-10-07 15:17:30.148323816 +0000 UTC m=+5532.231728141" lastFinishedPulling="2025-10-07 15:17:33.755000252 +0000 UTC m=+5535.838404577" observedRunningTime="2025-10-07 15:17:34.193971544 +0000 UTC m=+5536.277375879" watchObservedRunningTime="2025-10-07 15:17:34.196756635 +0000 UTC m=+5536.280160960" Oct 07 15:17:38 crc kubenswrapper[4959]: I1007 15:17:38.487987 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f2vz2" Oct 07 15:17:38 crc kubenswrapper[4959]: I1007 15:17:38.488591 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f2vz2" Oct 07 15:17:38 crc kubenswrapper[4959]: I1007 15:17:38.539226 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f2vz2" Oct 07 15:17:39 crc kubenswrapper[4959]: I1007 15:17:39.261409 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f2vz2" Oct 07 15:17:39 crc kubenswrapper[4959]: I1007 15:17:39.317705 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f2vz2"] Oct 07 15:17:41 crc kubenswrapper[4959]: I1007 15:17:41.232611 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-f2vz2" podUID="2bf22139-fc42-4993-97f7-0e1db00cc684" containerName="registry-server" containerID="cri-o://90e025308323baa19adb527cb7bc4a706e065a34ea9f2cbb981a7348b9d8cd82" gracePeriod=2 Oct 07 15:17:41 crc kubenswrapper[4959]: I1007 15:17:41.791845 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f2vz2" Oct 07 15:17:41 crc kubenswrapper[4959]: I1007 15:17:41.860251 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bf22139-fc42-4993-97f7-0e1db00cc684-catalog-content\") pod \"2bf22139-fc42-4993-97f7-0e1db00cc684\" (UID: \"2bf22139-fc42-4993-97f7-0e1db00cc684\") " Oct 07 15:17:41 crc kubenswrapper[4959]: I1007 15:17:41.860427 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8dpk\" (UniqueName: \"kubernetes.io/projected/2bf22139-fc42-4993-97f7-0e1db00cc684-kube-api-access-z8dpk\") pod \"2bf22139-fc42-4993-97f7-0e1db00cc684\" (UID: \"2bf22139-fc42-4993-97f7-0e1db00cc684\") " Oct 07 15:17:41 crc kubenswrapper[4959]: I1007 15:17:41.860496 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bf22139-fc42-4993-97f7-0e1db00cc684-utilities\") pod \"2bf22139-fc42-4993-97f7-0e1db00cc684\" (UID: \"2bf22139-fc42-4993-97f7-0e1db00cc684\") " Oct 07 15:17:41 crc kubenswrapper[4959]: I1007 15:17:41.861618 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bf22139-fc42-4993-97f7-0e1db00cc684-utilities" (OuterVolumeSpecName: "utilities") pod "2bf22139-fc42-4993-97f7-0e1db00cc684" (UID: "2bf22139-fc42-4993-97f7-0e1db00cc684"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:17:41 crc kubenswrapper[4959]: I1007 15:17:41.867954 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bf22139-fc42-4993-97f7-0e1db00cc684-kube-api-access-z8dpk" (OuterVolumeSpecName: "kube-api-access-z8dpk") pod "2bf22139-fc42-4993-97f7-0e1db00cc684" (UID: "2bf22139-fc42-4993-97f7-0e1db00cc684"). InnerVolumeSpecName "kube-api-access-z8dpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 15:17:41 crc kubenswrapper[4959]: I1007 15:17:41.962896 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8dpk\" (UniqueName: \"kubernetes.io/projected/2bf22139-fc42-4993-97f7-0e1db00cc684-kube-api-access-z8dpk\") on node \"crc\" DevicePath \"\"" Oct 07 15:17:41 crc kubenswrapper[4959]: I1007 15:17:41.962938 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bf22139-fc42-4993-97f7-0e1db00cc684-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 15:17:42 crc kubenswrapper[4959]: I1007 15:17:42.056934 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bf22139-fc42-4993-97f7-0e1db00cc684-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2bf22139-fc42-4993-97f7-0e1db00cc684" (UID: "2bf22139-fc42-4993-97f7-0e1db00cc684"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:17:42 crc kubenswrapper[4959]: I1007 15:17:42.065241 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bf22139-fc42-4993-97f7-0e1db00cc684-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 15:17:42 crc kubenswrapper[4959]: I1007 15:17:42.245946 4959 generic.go:334] "Generic (PLEG): container finished" podID="2bf22139-fc42-4993-97f7-0e1db00cc684" containerID="90e025308323baa19adb527cb7bc4a706e065a34ea9f2cbb981a7348b9d8cd82" exitCode=0 Oct 07 15:17:42 crc kubenswrapper[4959]: I1007 15:17:42.246001 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2vz2" event={"ID":"2bf22139-fc42-4993-97f7-0e1db00cc684","Type":"ContainerDied","Data":"90e025308323baa19adb527cb7bc4a706e065a34ea9f2cbb981a7348b9d8cd82"} Oct 07 15:17:42 crc kubenswrapper[4959]: I1007 15:17:42.246119 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f2vz2" Oct 07 15:17:42 crc kubenswrapper[4959]: I1007 15:17:42.246333 4959 scope.go:117] "RemoveContainer" containerID="90e025308323baa19adb527cb7bc4a706e065a34ea9f2cbb981a7348b9d8cd82" Oct 07 15:17:42 crc kubenswrapper[4959]: I1007 15:17:42.246314 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2vz2" event={"ID":"2bf22139-fc42-4993-97f7-0e1db00cc684","Type":"ContainerDied","Data":"c76c70a1c837be0093d01bddcd1816828113116d9d4ec92fea4ab28b458b276d"} Oct 07 15:17:42 crc kubenswrapper[4959]: I1007 15:17:42.281906 4959 scope.go:117] "RemoveContainer" containerID="28abefb82d22403588ad3046e81a173d36a10cca387cc444629e749bd9496c74" Oct 07 15:17:42 crc kubenswrapper[4959]: I1007 15:17:42.301272 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f2vz2"] Oct 07 15:17:42 crc kubenswrapper[4959]: I1007 15:17:42.310232 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-f2vz2"] Oct 07 15:17:42 crc kubenswrapper[4959]: I1007 15:17:42.315797 4959 scope.go:117] "RemoveContainer" containerID="b3d124a4ccdf6ce3f6af9f744f4663e49a9701341b22289f6ada766b66b3b74a" Oct 07 15:17:42 crc kubenswrapper[4959]: I1007 15:17:42.375729 4959 scope.go:117] "RemoveContainer" containerID="90e025308323baa19adb527cb7bc4a706e065a34ea9f2cbb981a7348b9d8cd82" Oct 07 15:17:42 crc kubenswrapper[4959]: E1007 15:17:42.376419 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90e025308323baa19adb527cb7bc4a706e065a34ea9f2cbb981a7348b9d8cd82\": container with ID starting with 90e025308323baa19adb527cb7bc4a706e065a34ea9f2cbb981a7348b9d8cd82 not found: ID does not exist" containerID="90e025308323baa19adb527cb7bc4a706e065a34ea9f2cbb981a7348b9d8cd82" Oct 07 15:17:42 crc kubenswrapper[4959]: I1007 15:17:42.376484 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90e025308323baa19adb527cb7bc4a706e065a34ea9f2cbb981a7348b9d8cd82"} err="failed to get container status \"90e025308323baa19adb527cb7bc4a706e065a34ea9f2cbb981a7348b9d8cd82\": rpc error: code = NotFound desc = could not find container \"90e025308323baa19adb527cb7bc4a706e065a34ea9f2cbb981a7348b9d8cd82\": container with ID starting with 90e025308323baa19adb527cb7bc4a706e065a34ea9f2cbb981a7348b9d8cd82 not found: ID does not exist" Oct 07 15:17:42 crc kubenswrapper[4959]: I1007 15:17:42.376515 4959 scope.go:117] "RemoveContainer" containerID="28abefb82d22403588ad3046e81a173d36a10cca387cc444629e749bd9496c74" Oct 07 15:17:42 crc kubenswrapper[4959]: E1007 15:17:42.376848 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28abefb82d22403588ad3046e81a173d36a10cca387cc444629e749bd9496c74\": container with ID starting with 28abefb82d22403588ad3046e81a173d36a10cca387cc444629e749bd9496c74 not found: ID does not exist" containerID="28abefb82d22403588ad3046e81a173d36a10cca387cc444629e749bd9496c74" Oct 07 15:17:42 crc kubenswrapper[4959]: I1007 15:17:42.376875 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28abefb82d22403588ad3046e81a173d36a10cca387cc444629e749bd9496c74"} err="failed to get container status \"28abefb82d22403588ad3046e81a173d36a10cca387cc444629e749bd9496c74\": rpc error: code = NotFound desc = could not find container \"28abefb82d22403588ad3046e81a173d36a10cca387cc444629e749bd9496c74\": container with ID starting with 28abefb82d22403588ad3046e81a173d36a10cca387cc444629e749bd9496c74 not found: ID does not exist" Oct 07 15:17:42 crc kubenswrapper[4959]: I1007 15:17:42.376893 4959 scope.go:117] "RemoveContainer" containerID="b3d124a4ccdf6ce3f6af9f744f4663e49a9701341b22289f6ada766b66b3b74a" Oct 07 15:17:42 crc kubenswrapper[4959]: E1007 15:17:42.377293 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3d124a4ccdf6ce3f6af9f744f4663e49a9701341b22289f6ada766b66b3b74a\": container with ID starting with b3d124a4ccdf6ce3f6af9f744f4663e49a9701341b22289f6ada766b66b3b74a not found: ID does not exist" containerID="b3d124a4ccdf6ce3f6af9f744f4663e49a9701341b22289f6ada766b66b3b74a" Oct 07 15:17:42 crc kubenswrapper[4959]: I1007 15:17:42.377323 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3d124a4ccdf6ce3f6af9f744f4663e49a9701341b22289f6ada766b66b3b74a"} err="failed to get container status \"b3d124a4ccdf6ce3f6af9f744f4663e49a9701341b22289f6ada766b66b3b74a\": rpc error: code = NotFound desc = could not find container \"b3d124a4ccdf6ce3f6af9f744f4663e49a9701341b22289f6ada766b66b3b74a\": container with ID starting with b3d124a4ccdf6ce3f6af9f744f4663e49a9701341b22289f6ada766b66b3b74a not found: ID does not exist" Oct 07 15:17:42 crc kubenswrapper[4959]: I1007 15:17:42.667375 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bf22139-fc42-4993-97f7-0e1db00cc684" path="/var/lib/kubelet/pods/2bf22139-fc42-4993-97f7-0e1db00cc684/volumes" Oct 07 15:17:50 crc kubenswrapper[4959]: I1007 15:17:50.632940 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 15:17:50 crc kubenswrapper[4959]: I1007 15:17:50.635482 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 15:17:50 crc kubenswrapper[4959]: I1007 15:17:50.635649 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 15:17:50 crc kubenswrapper[4959]: I1007 15:17:50.637192 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2c46cfa18edd8dc1249072876738f373407f694f7fb80d8b963bc7e73086ff89"} pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 15:17:50 crc kubenswrapper[4959]: I1007 15:17:50.637364 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" containerID="cri-o://2c46cfa18edd8dc1249072876738f373407f694f7fb80d8b963bc7e73086ff89" gracePeriod=600 Oct 07 15:17:51 crc kubenswrapper[4959]: I1007 15:17:51.376286 4959 generic.go:334] "Generic (PLEG): container finished" podID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerID="2c46cfa18edd8dc1249072876738f373407f694f7fb80d8b963bc7e73086ff89" exitCode=0 Oct 07 15:17:51 crc kubenswrapper[4959]: I1007 15:17:51.376396 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerDied","Data":"2c46cfa18edd8dc1249072876738f373407f694f7fb80d8b963bc7e73086ff89"} Oct 07 15:17:51 crc kubenswrapper[4959]: I1007 15:17:51.378424 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerStarted","Data":"20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168"} Oct 07 15:17:51 crc kubenswrapper[4959]: I1007 15:17:51.378472 4959 scope.go:117] "RemoveContainer" containerID="e3c076ba1b4a9c0a71ed69ece9a2159844ff22445cf7fc8a46dad6e4b29e3074" Oct 07 15:17:51 crc kubenswrapper[4959]: I1007 15:17:51.505481 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jcw8q"] Oct 07 15:17:51 crc kubenswrapper[4959]: E1007 15:17:51.509452 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bf22139-fc42-4993-97f7-0e1db00cc684" containerName="extract-content" Oct 07 15:17:51 crc kubenswrapper[4959]: I1007 15:17:51.509479 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bf22139-fc42-4993-97f7-0e1db00cc684" containerName="extract-content" Oct 07 15:17:51 crc kubenswrapper[4959]: E1007 15:17:51.509731 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bf22139-fc42-4993-97f7-0e1db00cc684" containerName="registry-server" Oct 07 15:17:51 crc kubenswrapper[4959]: I1007 15:17:51.509860 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bf22139-fc42-4993-97f7-0e1db00cc684" containerName="registry-server" Oct 07 15:17:51 crc kubenswrapper[4959]: E1007 15:17:51.510068 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bf22139-fc42-4993-97f7-0e1db00cc684" containerName="extract-utilities" Oct 07 15:17:51 crc kubenswrapper[4959]: I1007 15:17:51.510081 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bf22139-fc42-4993-97f7-0e1db00cc684" containerName="extract-utilities" Oct 07 15:17:51 crc kubenswrapper[4959]: I1007 15:17:51.514256 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bf22139-fc42-4993-97f7-0e1db00cc684" containerName="registry-server" Oct 07 15:17:51 crc kubenswrapper[4959]: I1007 15:17:51.531124 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jcw8q" Oct 07 15:17:51 crc kubenswrapper[4959]: I1007 15:17:51.545709 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jcw8q"] Oct 07 15:17:51 crc kubenswrapper[4959]: I1007 15:17:51.589378 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms7cp\" (UniqueName: \"kubernetes.io/projected/cd73339f-801a-4cd7-b871-9f16c4781a42-kube-api-access-ms7cp\") pod \"redhat-marketplace-jcw8q\" (UID: \"cd73339f-801a-4cd7-b871-9f16c4781a42\") " pod="openshift-marketplace/redhat-marketplace-jcw8q" Oct 07 15:17:51 crc kubenswrapper[4959]: I1007 15:17:51.589465 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd73339f-801a-4cd7-b871-9f16c4781a42-catalog-content\") pod \"redhat-marketplace-jcw8q\" (UID: \"cd73339f-801a-4cd7-b871-9f16c4781a42\") " pod="openshift-marketplace/redhat-marketplace-jcw8q" Oct 07 15:17:51 crc kubenswrapper[4959]: I1007 15:17:51.589511 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd73339f-801a-4cd7-b871-9f16c4781a42-utilities\") pod \"redhat-marketplace-jcw8q\" (UID: \"cd73339f-801a-4cd7-b871-9f16c4781a42\") " pod="openshift-marketplace/redhat-marketplace-jcw8q" Oct 07 15:17:51 crc kubenswrapper[4959]: I1007 15:17:51.693293 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms7cp\" (UniqueName: \"kubernetes.io/projected/cd73339f-801a-4cd7-b871-9f16c4781a42-kube-api-access-ms7cp\") pod \"redhat-marketplace-jcw8q\" (UID: \"cd73339f-801a-4cd7-b871-9f16c4781a42\") " pod="openshift-marketplace/redhat-marketplace-jcw8q" Oct 07 15:17:51 crc kubenswrapper[4959]: I1007 15:17:51.693419 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd73339f-801a-4cd7-b871-9f16c4781a42-catalog-content\") pod \"redhat-marketplace-jcw8q\" (UID: \"cd73339f-801a-4cd7-b871-9f16c4781a42\") " pod="openshift-marketplace/redhat-marketplace-jcw8q" Oct 07 15:17:51 crc kubenswrapper[4959]: I1007 15:17:51.693449 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd73339f-801a-4cd7-b871-9f16c4781a42-utilities\") pod \"redhat-marketplace-jcw8q\" (UID: \"cd73339f-801a-4cd7-b871-9f16c4781a42\") " pod="openshift-marketplace/redhat-marketplace-jcw8q" Oct 07 15:17:51 crc kubenswrapper[4959]: I1007 15:17:51.694258 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd73339f-801a-4cd7-b871-9f16c4781a42-utilities\") pod \"redhat-marketplace-jcw8q\" (UID: \"cd73339f-801a-4cd7-b871-9f16c4781a42\") " pod="openshift-marketplace/redhat-marketplace-jcw8q" Oct 07 15:17:51 crc kubenswrapper[4959]: I1007 15:17:51.694669 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd73339f-801a-4cd7-b871-9f16c4781a42-catalog-content\") pod \"redhat-marketplace-jcw8q\" (UID: \"cd73339f-801a-4cd7-b871-9f16c4781a42\") " pod="openshift-marketplace/redhat-marketplace-jcw8q" Oct 07 15:17:51 crc kubenswrapper[4959]: I1007 15:17:51.723547 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms7cp\" (UniqueName: \"kubernetes.io/projected/cd73339f-801a-4cd7-b871-9f16c4781a42-kube-api-access-ms7cp\") pod \"redhat-marketplace-jcw8q\" (UID: \"cd73339f-801a-4cd7-b871-9f16c4781a42\") " pod="openshift-marketplace/redhat-marketplace-jcw8q" Oct 07 15:17:51 crc kubenswrapper[4959]: I1007 15:17:51.862774 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jcw8q" Oct 07 15:17:52 crc kubenswrapper[4959]: I1007 15:17:52.398253 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jcw8q"] Oct 07 15:17:53 crc kubenswrapper[4959]: I1007 15:17:53.407828 4959 generic.go:334] "Generic (PLEG): container finished" podID="cd73339f-801a-4cd7-b871-9f16c4781a42" containerID="3a5698f618e41a0f3ec535c0a136fdf2ece0c098122b7fcd5316e1d7823553ed" exitCode=0 Oct 07 15:17:53 crc kubenswrapper[4959]: I1007 15:17:53.407911 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jcw8q" event={"ID":"cd73339f-801a-4cd7-b871-9f16c4781a42","Type":"ContainerDied","Data":"3a5698f618e41a0f3ec535c0a136fdf2ece0c098122b7fcd5316e1d7823553ed"} Oct 07 15:17:53 crc kubenswrapper[4959]: I1007 15:17:53.408333 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jcw8q" event={"ID":"cd73339f-801a-4cd7-b871-9f16c4781a42","Type":"ContainerStarted","Data":"42d3e41a2710b5553e1f0abb60c7c41b3b9951b50b3b973ca324ec9a8a03e15f"} Oct 07 15:17:53 crc kubenswrapper[4959]: I1007 15:17:53.410761 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 15:17:54 crc kubenswrapper[4959]: I1007 15:17:54.422334 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jcw8q" event={"ID":"cd73339f-801a-4cd7-b871-9f16c4781a42","Type":"ContainerStarted","Data":"7b59a462d9ba09af902ee21986d0e8cf9b0e1dd106b305472beea5c8e7e5e8f9"} Oct 07 15:17:55 crc kubenswrapper[4959]: I1007 15:17:55.436602 4959 generic.go:334] "Generic (PLEG): container finished" podID="cd73339f-801a-4cd7-b871-9f16c4781a42" containerID="7b59a462d9ba09af902ee21986d0e8cf9b0e1dd106b305472beea5c8e7e5e8f9" exitCode=0 Oct 07 15:17:55 crc kubenswrapper[4959]: I1007 15:17:55.436728 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jcw8q" event={"ID":"cd73339f-801a-4cd7-b871-9f16c4781a42","Type":"ContainerDied","Data":"7b59a462d9ba09af902ee21986d0e8cf9b0e1dd106b305472beea5c8e7e5e8f9"} Oct 07 15:17:56 crc kubenswrapper[4959]: I1007 15:17:56.468817 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jcw8q" event={"ID":"cd73339f-801a-4cd7-b871-9f16c4781a42","Type":"ContainerStarted","Data":"a6f4b0e2cfafdc00709bfca6cd42bd3c3d58877a6f011db56ab62129b4ce780f"} Oct 07 15:17:56 crc kubenswrapper[4959]: I1007 15:17:56.500006 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jcw8q" podStartSLOduration=2.989963671 podStartE2EDuration="5.49998126s" podCreationTimestamp="2025-10-07 15:17:51 +0000 UTC" firstStartedPulling="2025-10-07 15:17:53.410464687 +0000 UTC m=+5555.493869012" lastFinishedPulling="2025-10-07 15:17:55.920482276 +0000 UTC m=+5558.003886601" observedRunningTime="2025-10-07 15:17:56.487702936 +0000 UTC m=+5558.571107271" watchObservedRunningTime="2025-10-07 15:17:56.49998126 +0000 UTC m=+5558.583385585" Oct 07 15:18:01 crc kubenswrapper[4959]: I1007 15:18:01.863033 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jcw8q" Oct 07 15:18:01 crc kubenswrapper[4959]: I1007 15:18:01.863625 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jcw8q" Oct 07 15:18:01 crc kubenswrapper[4959]: I1007 15:18:01.933282 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jcw8q" Oct 07 15:18:02 crc kubenswrapper[4959]: I1007 15:18:02.587996 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jcw8q" Oct 07 15:18:02 crc kubenswrapper[4959]: I1007 15:18:02.667735 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jcw8q"] Oct 07 15:18:04 crc kubenswrapper[4959]: I1007 15:18:04.549747 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jcw8q" podUID="cd73339f-801a-4cd7-b871-9f16c4781a42" containerName="registry-server" containerID="cri-o://a6f4b0e2cfafdc00709bfca6cd42bd3c3d58877a6f011db56ab62129b4ce780f" gracePeriod=2 Oct 07 15:18:05 crc kubenswrapper[4959]: I1007 15:18:05.132560 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jcw8q" Oct 07 15:18:05 crc kubenswrapper[4959]: I1007 15:18:05.302935 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd73339f-801a-4cd7-b871-9f16c4781a42-utilities\") pod \"cd73339f-801a-4cd7-b871-9f16c4781a42\" (UID: \"cd73339f-801a-4cd7-b871-9f16c4781a42\") " Oct 07 15:18:05 crc kubenswrapper[4959]: I1007 15:18:05.303011 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ms7cp\" (UniqueName: \"kubernetes.io/projected/cd73339f-801a-4cd7-b871-9f16c4781a42-kube-api-access-ms7cp\") pod \"cd73339f-801a-4cd7-b871-9f16c4781a42\" (UID: \"cd73339f-801a-4cd7-b871-9f16c4781a42\") " Oct 07 15:18:05 crc kubenswrapper[4959]: I1007 15:18:05.303273 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd73339f-801a-4cd7-b871-9f16c4781a42-catalog-content\") pod \"cd73339f-801a-4cd7-b871-9f16c4781a42\" (UID: \"cd73339f-801a-4cd7-b871-9f16c4781a42\") " Oct 07 15:18:05 crc kubenswrapper[4959]: I1007 15:18:05.304351 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd73339f-801a-4cd7-b871-9f16c4781a42-utilities" (OuterVolumeSpecName: "utilities") pod "cd73339f-801a-4cd7-b871-9f16c4781a42" (UID: "cd73339f-801a-4cd7-b871-9f16c4781a42"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:18:05 crc kubenswrapper[4959]: I1007 15:18:05.312050 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd73339f-801a-4cd7-b871-9f16c4781a42-kube-api-access-ms7cp" (OuterVolumeSpecName: "kube-api-access-ms7cp") pod "cd73339f-801a-4cd7-b871-9f16c4781a42" (UID: "cd73339f-801a-4cd7-b871-9f16c4781a42"). InnerVolumeSpecName "kube-api-access-ms7cp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 15:18:05 crc kubenswrapper[4959]: I1007 15:18:05.317918 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd73339f-801a-4cd7-b871-9f16c4781a42-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cd73339f-801a-4cd7-b871-9f16c4781a42" (UID: "cd73339f-801a-4cd7-b871-9f16c4781a42"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:18:05 crc kubenswrapper[4959]: I1007 15:18:05.407192 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd73339f-801a-4cd7-b871-9f16c4781a42-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 15:18:05 crc kubenswrapper[4959]: I1007 15:18:05.407245 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd73339f-801a-4cd7-b871-9f16c4781a42-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 15:18:05 crc kubenswrapper[4959]: I1007 15:18:05.407255 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ms7cp\" (UniqueName: \"kubernetes.io/projected/cd73339f-801a-4cd7-b871-9f16c4781a42-kube-api-access-ms7cp\") on node \"crc\" DevicePath \"\"" Oct 07 15:18:05 crc kubenswrapper[4959]: I1007 15:18:05.562337 4959 generic.go:334] "Generic (PLEG): container finished" podID="cd73339f-801a-4cd7-b871-9f16c4781a42" containerID="a6f4b0e2cfafdc00709bfca6cd42bd3c3d58877a6f011db56ab62129b4ce780f" exitCode=0 Oct 07 15:18:05 crc kubenswrapper[4959]: I1007 15:18:05.562401 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jcw8q" Oct 07 15:18:05 crc kubenswrapper[4959]: I1007 15:18:05.562397 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jcw8q" event={"ID":"cd73339f-801a-4cd7-b871-9f16c4781a42","Type":"ContainerDied","Data":"a6f4b0e2cfafdc00709bfca6cd42bd3c3d58877a6f011db56ab62129b4ce780f"} Oct 07 15:18:05 crc kubenswrapper[4959]: I1007 15:18:05.562632 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jcw8q" event={"ID":"cd73339f-801a-4cd7-b871-9f16c4781a42","Type":"ContainerDied","Data":"42d3e41a2710b5553e1f0abb60c7c41b3b9951b50b3b973ca324ec9a8a03e15f"} Oct 07 15:18:05 crc kubenswrapper[4959]: I1007 15:18:05.562665 4959 scope.go:117] "RemoveContainer" containerID="a6f4b0e2cfafdc00709bfca6cd42bd3c3d58877a6f011db56ab62129b4ce780f" Oct 07 15:18:05 crc kubenswrapper[4959]: I1007 15:18:05.598983 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jcw8q"] Oct 07 15:18:05 crc kubenswrapper[4959]: I1007 15:18:05.600829 4959 scope.go:117] "RemoveContainer" containerID="7b59a462d9ba09af902ee21986d0e8cf9b0e1dd106b305472beea5c8e7e5e8f9" Oct 07 15:18:05 crc kubenswrapper[4959]: I1007 15:18:05.607646 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jcw8q"] Oct 07 15:18:05 crc kubenswrapper[4959]: I1007 15:18:05.635659 4959 scope.go:117] "RemoveContainer" containerID="3a5698f618e41a0f3ec535c0a136fdf2ece0c098122b7fcd5316e1d7823553ed" Oct 07 15:18:05 crc kubenswrapper[4959]: I1007 15:18:05.685755 4959 scope.go:117] "RemoveContainer" containerID="a6f4b0e2cfafdc00709bfca6cd42bd3c3d58877a6f011db56ab62129b4ce780f" Oct 07 15:18:05 crc kubenswrapper[4959]: E1007 15:18:05.686997 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6f4b0e2cfafdc00709bfca6cd42bd3c3d58877a6f011db56ab62129b4ce780f\": container with ID starting with a6f4b0e2cfafdc00709bfca6cd42bd3c3d58877a6f011db56ab62129b4ce780f not found: ID does not exist" containerID="a6f4b0e2cfafdc00709bfca6cd42bd3c3d58877a6f011db56ab62129b4ce780f" Oct 07 15:18:05 crc kubenswrapper[4959]: I1007 15:18:05.687064 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6f4b0e2cfafdc00709bfca6cd42bd3c3d58877a6f011db56ab62129b4ce780f"} err="failed to get container status \"a6f4b0e2cfafdc00709bfca6cd42bd3c3d58877a6f011db56ab62129b4ce780f\": rpc error: code = NotFound desc = could not find container \"a6f4b0e2cfafdc00709bfca6cd42bd3c3d58877a6f011db56ab62129b4ce780f\": container with ID starting with a6f4b0e2cfafdc00709bfca6cd42bd3c3d58877a6f011db56ab62129b4ce780f not found: ID does not exist" Oct 07 15:18:05 crc kubenswrapper[4959]: I1007 15:18:05.687166 4959 scope.go:117] "RemoveContainer" containerID="7b59a462d9ba09af902ee21986d0e8cf9b0e1dd106b305472beea5c8e7e5e8f9" Oct 07 15:18:05 crc kubenswrapper[4959]: E1007 15:18:05.687661 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b59a462d9ba09af902ee21986d0e8cf9b0e1dd106b305472beea5c8e7e5e8f9\": container with ID starting with 7b59a462d9ba09af902ee21986d0e8cf9b0e1dd106b305472beea5c8e7e5e8f9 not found: ID does not exist" containerID="7b59a462d9ba09af902ee21986d0e8cf9b0e1dd106b305472beea5c8e7e5e8f9" Oct 07 15:18:05 crc kubenswrapper[4959]: I1007 15:18:05.687702 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b59a462d9ba09af902ee21986d0e8cf9b0e1dd106b305472beea5c8e7e5e8f9"} err="failed to get container status \"7b59a462d9ba09af902ee21986d0e8cf9b0e1dd106b305472beea5c8e7e5e8f9\": rpc error: code = NotFound desc = could not find container \"7b59a462d9ba09af902ee21986d0e8cf9b0e1dd106b305472beea5c8e7e5e8f9\": container with ID starting with 7b59a462d9ba09af902ee21986d0e8cf9b0e1dd106b305472beea5c8e7e5e8f9 not found: ID does not exist" Oct 07 15:18:05 crc kubenswrapper[4959]: I1007 15:18:05.687733 4959 scope.go:117] "RemoveContainer" containerID="3a5698f618e41a0f3ec535c0a136fdf2ece0c098122b7fcd5316e1d7823553ed" Oct 07 15:18:05 crc kubenswrapper[4959]: E1007 15:18:05.688212 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a5698f618e41a0f3ec535c0a136fdf2ece0c098122b7fcd5316e1d7823553ed\": container with ID starting with 3a5698f618e41a0f3ec535c0a136fdf2ece0c098122b7fcd5316e1d7823553ed not found: ID does not exist" containerID="3a5698f618e41a0f3ec535c0a136fdf2ece0c098122b7fcd5316e1d7823553ed" Oct 07 15:18:05 crc kubenswrapper[4959]: I1007 15:18:05.688236 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a5698f618e41a0f3ec535c0a136fdf2ece0c098122b7fcd5316e1d7823553ed"} err="failed to get container status \"3a5698f618e41a0f3ec535c0a136fdf2ece0c098122b7fcd5316e1d7823553ed\": rpc error: code = NotFound desc = could not find container \"3a5698f618e41a0f3ec535c0a136fdf2ece0c098122b7fcd5316e1d7823553ed\": container with ID starting with 3a5698f618e41a0f3ec535c0a136fdf2ece0c098122b7fcd5316e1d7823553ed not found: ID does not exist" Oct 07 15:18:06 crc kubenswrapper[4959]: I1007 15:18:06.664765 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd73339f-801a-4cd7-b871-9f16c4781a42" path="/var/lib/kubelet/pods/cd73339f-801a-4cd7-b871-9f16c4781a42/volumes" Oct 07 15:20:20 crc kubenswrapper[4959]: I1007 15:20:20.630750 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 15:20:20 crc kubenswrapper[4959]: I1007 15:20:20.631803 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 15:20:50 crc kubenswrapper[4959]: I1007 15:20:50.630766 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 15:20:50 crc kubenswrapper[4959]: I1007 15:20:50.631337 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 15:21:20 crc kubenswrapper[4959]: I1007 15:21:20.630954 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 15:21:20 crc kubenswrapper[4959]: I1007 15:21:20.631955 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 15:21:20 crc kubenswrapper[4959]: I1007 15:21:20.632041 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 15:21:20 crc kubenswrapper[4959]: I1007 15:21:20.633433 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168"} pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 15:21:20 crc kubenswrapper[4959]: I1007 15:21:20.633516 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" containerID="cri-o://20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168" gracePeriod=600 Oct 07 15:21:20 crc kubenswrapper[4959]: E1007 15:21:20.760670 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:21:21 crc kubenswrapper[4959]: I1007 15:21:21.457611 4959 generic.go:334] "Generic (PLEG): container finished" podID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerID="20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168" exitCode=0 Oct 07 15:21:21 crc kubenswrapper[4959]: I1007 15:21:21.457672 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerDied","Data":"20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168"} Oct 07 15:21:21 crc kubenswrapper[4959]: I1007 15:21:21.457719 4959 scope.go:117] "RemoveContainer" containerID="2c46cfa18edd8dc1249072876738f373407f694f7fb80d8b963bc7e73086ff89" Oct 07 15:21:21 crc kubenswrapper[4959]: I1007 15:21:21.458563 4959 scope.go:117] "RemoveContainer" containerID="20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168" Oct 07 15:21:21 crc kubenswrapper[4959]: E1007 15:21:21.458912 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:21:36 crc kubenswrapper[4959]: I1007 15:21:36.655616 4959 scope.go:117] "RemoveContainer" containerID="20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168" Oct 07 15:21:36 crc kubenswrapper[4959]: E1007 15:21:36.656507 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:21:48 crc kubenswrapper[4959]: I1007 15:21:48.661234 4959 scope.go:117] "RemoveContainer" containerID="20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168" Oct 07 15:21:48 crc kubenswrapper[4959]: E1007 15:21:48.662280 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:22:00 crc kubenswrapper[4959]: I1007 15:22:00.852713 4959 generic.go:334] "Generic (PLEG): container finished" podID="b68f21fe-48e8-4b78-8476-ec53a2ca30c8" containerID="7450260130b50a2b82b1c1e913dc57c1afa0f0164a7a1a01cdb7cf5f266fb493" exitCode=1 Oct 07 15:22:00 crc kubenswrapper[4959]: I1007 15:22:00.852838 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-full" event={"ID":"b68f21fe-48e8-4b78-8476-ec53a2ca30c8","Type":"ContainerDied","Data":"7450260130b50a2b82b1c1e913dc57c1afa0f0164a7a1a01cdb7cf5f266fb493"} Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.372152 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-full" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.463165 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest-s01-single-test"] Oct 07 15:22:02 crc kubenswrapper[4959]: E1007 15:22:02.463674 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd73339f-801a-4cd7-b871-9f16c4781a42" containerName="extract-content" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.463693 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd73339f-801a-4cd7-b871-9f16c4781a42" containerName="extract-content" Oct 07 15:22:02 crc kubenswrapper[4959]: E1007 15:22:02.463709 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd73339f-801a-4cd7-b871-9f16c4781a42" containerName="registry-server" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.463715 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd73339f-801a-4cd7-b871-9f16c4781a42" containerName="registry-server" Oct 07 15:22:02 crc kubenswrapper[4959]: E1007 15:22:02.463732 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b68f21fe-48e8-4b78-8476-ec53a2ca30c8" containerName="tempest-tests-tempest-tests-runner" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.463737 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b68f21fe-48e8-4b78-8476-ec53a2ca30c8" containerName="tempest-tests-tempest-tests-runner" Oct 07 15:22:02 crc kubenswrapper[4959]: E1007 15:22:02.463748 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd73339f-801a-4cd7-b871-9f16c4781a42" containerName="extract-utilities" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.463754 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd73339f-801a-4cd7-b871-9f16c4781a42" containerName="extract-utilities" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.463936 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd73339f-801a-4cd7-b871-9f16c4781a42" containerName="registry-server" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.463960 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b68f21fe-48e8-4b78-8476-ec53a2ca30c8" containerName="tempest-tests-tempest-tests-runner" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.464739 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.468411 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s1" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.468411 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s1" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.470351 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-openstack-config-secret\") pod \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.470479 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-ceph\") pod \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.470526 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-ca-certs\") pod \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.470563 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-openstack-config\") pod \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.470613 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.470702 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74sqm\" (UniqueName: \"kubernetes.io/projected/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-kube-api-access-74sqm\") pod \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.470776 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-ssh-key\") pod \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.470844 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-test-operator-ephemeral-workdir\") pod \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.470872 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-test-operator-ephemeral-temporary\") pod \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.470896 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-config-data\") pod \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\" (UID: \"b68f21fe-48e8-4b78-8476-ec53a2ca30c8\") " Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.471547 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s01-single-test"] Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.472832 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "b68f21fe-48e8-4b78-8476-ec53a2ca30c8" (UID: "b68f21fe-48e8-4b78-8476-ec53a2ca30c8"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.480269 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-config-data" (OuterVolumeSpecName: "config-data") pod "b68f21fe-48e8-4b78-8476-ec53a2ca30c8" (UID: "b68f21fe-48e8-4b78-8476-ec53a2ca30c8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.480410 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "test-operator-logs") pod "b68f21fe-48e8-4b78-8476-ec53a2ca30c8" (UID: "b68f21fe-48e8-4b78-8476-ec53a2ca30c8"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.481075 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-kube-api-access-74sqm" (OuterVolumeSpecName: "kube-api-access-74sqm") pod "b68f21fe-48e8-4b78-8476-ec53a2ca30c8" (UID: "b68f21fe-48e8-4b78-8476-ec53a2ca30c8"). InnerVolumeSpecName "kube-api-access-74sqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.483075 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "b68f21fe-48e8-4b78-8476-ec53a2ca30c8" (UID: "b68f21fe-48e8-4b78-8476-ec53a2ca30c8"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.508253 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-ceph" (OuterVolumeSpecName: "ceph") pod "b68f21fe-48e8-4b78-8476-ec53a2ca30c8" (UID: "b68f21fe-48e8-4b78-8476-ec53a2ca30c8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.526295 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b68f21fe-48e8-4b78-8476-ec53a2ca30c8" (UID: "b68f21fe-48e8-4b78-8476-ec53a2ca30c8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.535213 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "b68f21fe-48e8-4b78-8476-ec53a2ca30c8" (UID: "b68f21fe-48e8-4b78-8476-ec53a2ca30c8"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.538599 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "b68f21fe-48e8-4b78-8476-ec53a2ca30c8" (UID: "b68f21fe-48e8-4b78-8476-ec53a2ca30c8"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.562477 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "b68f21fe-48e8-4b78-8476-ec53a2ca30c8" (UID: "b68f21fe-48e8-4b78-8476-ec53a2ca30c8"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.572591 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.572655 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-ceph\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.572715 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.572803 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.572853 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-ca-certs\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.572882 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-ssh-key\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.572936 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-config-data\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.572967 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.572993 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n89v8\" (UniqueName: \"kubernetes.io/projected/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-kube-api-access-n89v8\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.573044 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-openstack-config\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.573154 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74sqm\" (UniqueName: \"kubernetes.io/projected/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-kube-api-access-74sqm\") on node \"crc\" DevicePath \"\"" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.573167 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.573178 4959 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.573188 4959 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.573198 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.573209 4959 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.573224 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-ceph\") on node \"crc\" DevicePath \"\"" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.573239 4959 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.573252 4959 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b68f21fe-48e8-4b78-8476-ec53a2ca30c8-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.605819 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.674154 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-openstack-config\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.674279 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.674310 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-ceph\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.674345 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.674452 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-ca-certs\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.674494 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-ssh-key\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.674564 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-config-data\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.674587 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.674603 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n89v8\" (UniqueName: \"kubernetes.io/projected/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-kube-api-access-n89v8\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.679189 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.679297 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.679464 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-openstack-config\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.680204 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.680232 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-config-data\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.681793 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-ssh-key\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.684688 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-ceph\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.685045 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-ca-certs\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.692963 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n89v8\" (UniqueName: \"kubernetes.io/projected/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-kube-api-access-n89v8\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.891090 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-full" event={"ID":"b68f21fe-48e8-4b78-8476-ec53a2ca30c8","Type":"ContainerDied","Data":"c0c64a8bd6af2159fb78c62dce1c6e94ac62e28112e7112d82445a67061eab65"} Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.891189 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0c64a8bd6af2159fb78c62dce1c6e94ac62e28112e7112d82445a67061eab65" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.891260 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-full" Oct 07 15:22:02 crc kubenswrapper[4959]: I1007 15:22:02.951711 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:22:03 crc kubenswrapper[4959]: I1007 15:22:03.640447 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s01-single-test"] Oct 07 15:22:03 crc kubenswrapper[4959]: I1007 15:22:03.654482 4959 scope.go:117] "RemoveContainer" containerID="20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168" Oct 07 15:22:03 crc kubenswrapper[4959]: E1007 15:22:03.654845 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:22:03 crc kubenswrapper[4959]: I1007 15:22:03.906630 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-test" event={"ID":"bc180a96-fac6-4b97-97bf-11b7c7ceff8a","Type":"ContainerStarted","Data":"e23f8218120916894a8f6c03f741cc196294af851ffa4762217348e4c312191e"} Oct 07 15:22:04 crc kubenswrapper[4959]: I1007 15:22:04.918634 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-test" event={"ID":"bc180a96-fac6-4b97-97bf-11b7c7ceff8a","Type":"ContainerStarted","Data":"83923ad3156f01c3fd190fb9bddab0839d9f7a15acf5197b50ba35bd871a4fe5"} Oct 07 15:22:04 crc kubenswrapper[4959]: I1007 15:22:04.940638 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest-s01-single-test" podStartSLOduration=2.940575541 podStartE2EDuration="2.940575541s" podCreationTimestamp="2025-10-07 15:22:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 15:22:04.935337957 +0000 UTC m=+5807.018742282" watchObservedRunningTime="2025-10-07 15:22:04.940575541 +0000 UTC m=+5807.023979866" Oct 07 15:22:15 crc kubenswrapper[4959]: I1007 15:22:15.654134 4959 scope.go:117] "RemoveContainer" containerID="20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168" Oct 07 15:22:15 crc kubenswrapper[4959]: E1007 15:22:15.655035 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:22:30 crc kubenswrapper[4959]: I1007 15:22:30.654620 4959 scope.go:117] "RemoveContainer" containerID="20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168" Oct 07 15:22:30 crc kubenswrapper[4959]: E1007 15:22:30.655729 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:22:44 crc kubenswrapper[4959]: I1007 15:22:44.654586 4959 scope.go:117] "RemoveContainer" containerID="20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168" Oct 07 15:22:44 crc kubenswrapper[4959]: E1007 15:22:44.655728 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:22:55 crc kubenswrapper[4959]: I1007 15:22:55.653976 4959 scope.go:117] "RemoveContainer" containerID="20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168" Oct 07 15:22:55 crc kubenswrapper[4959]: E1007 15:22:55.657122 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:23:08 crc kubenswrapper[4959]: I1007 15:23:08.661710 4959 scope.go:117] "RemoveContainer" containerID="20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168" Oct 07 15:23:08 crc kubenswrapper[4959]: E1007 15:23:08.663157 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:23:23 crc kubenswrapper[4959]: I1007 15:23:23.653560 4959 scope.go:117] "RemoveContainer" containerID="20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168" Oct 07 15:23:23 crc kubenswrapper[4959]: E1007 15:23:23.654519 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:23:37 crc kubenswrapper[4959]: I1007 15:23:37.654392 4959 scope.go:117] "RemoveContainer" containerID="20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168" Oct 07 15:23:37 crc kubenswrapper[4959]: E1007 15:23:37.655582 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:23:49 crc kubenswrapper[4959]: I1007 15:23:49.654286 4959 scope.go:117] "RemoveContainer" containerID="20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168" Oct 07 15:23:49 crc kubenswrapper[4959]: E1007 15:23:49.655799 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:23:56 crc kubenswrapper[4959]: I1007 15:23:56.052418 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lk5mc"] Oct 07 15:23:56 crc kubenswrapper[4959]: I1007 15:23:56.054985 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lk5mc" Oct 07 15:23:56 crc kubenswrapper[4959]: I1007 15:23:56.069619 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lk5mc"] Oct 07 15:23:56 crc kubenswrapper[4959]: I1007 15:23:56.114634 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmrmg\" (UniqueName: \"kubernetes.io/projected/c34c5d88-8426-4e1c-a547-4688cb55e2ae-kube-api-access-nmrmg\") pod \"certified-operators-lk5mc\" (UID: \"c34c5d88-8426-4e1c-a547-4688cb55e2ae\") " pod="openshift-marketplace/certified-operators-lk5mc" Oct 07 15:23:56 crc kubenswrapper[4959]: I1007 15:23:56.114699 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c34c5d88-8426-4e1c-a547-4688cb55e2ae-utilities\") pod \"certified-operators-lk5mc\" (UID: \"c34c5d88-8426-4e1c-a547-4688cb55e2ae\") " pod="openshift-marketplace/certified-operators-lk5mc" Oct 07 15:23:56 crc kubenswrapper[4959]: I1007 15:23:56.115278 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c34c5d88-8426-4e1c-a547-4688cb55e2ae-catalog-content\") pod \"certified-operators-lk5mc\" (UID: \"c34c5d88-8426-4e1c-a547-4688cb55e2ae\") " pod="openshift-marketplace/certified-operators-lk5mc" Oct 07 15:23:56 crc kubenswrapper[4959]: I1007 15:23:56.216930 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c34c5d88-8426-4e1c-a547-4688cb55e2ae-catalog-content\") pod \"certified-operators-lk5mc\" (UID: \"c34c5d88-8426-4e1c-a547-4688cb55e2ae\") " pod="openshift-marketplace/certified-operators-lk5mc" Oct 07 15:23:56 crc kubenswrapper[4959]: I1007 15:23:56.217043 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmrmg\" (UniqueName: \"kubernetes.io/projected/c34c5d88-8426-4e1c-a547-4688cb55e2ae-kube-api-access-nmrmg\") pod \"certified-operators-lk5mc\" (UID: \"c34c5d88-8426-4e1c-a547-4688cb55e2ae\") " pod="openshift-marketplace/certified-operators-lk5mc" Oct 07 15:23:56 crc kubenswrapper[4959]: I1007 15:23:56.217075 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c34c5d88-8426-4e1c-a547-4688cb55e2ae-utilities\") pod \"certified-operators-lk5mc\" (UID: \"c34c5d88-8426-4e1c-a547-4688cb55e2ae\") " pod="openshift-marketplace/certified-operators-lk5mc" Oct 07 15:23:56 crc kubenswrapper[4959]: I1007 15:23:56.217500 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c34c5d88-8426-4e1c-a547-4688cb55e2ae-catalog-content\") pod \"certified-operators-lk5mc\" (UID: \"c34c5d88-8426-4e1c-a547-4688cb55e2ae\") " pod="openshift-marketplace/certified-operators-lk5mc" Oct 07 15:23:56 crc kubenswrapper[4959]: I1007 15:23:56.217634 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c34c5d88-8426-4e1c-a547-4688cb55e2ae-utilities\") pod \"certified-operators-lk5mc\" (UID: \"c34c5d88-8426-4e1c-a547-4688cb55e2ae\") " pod="openshift-marketplace/certified-operators-lk5mc" Oct 07 15:23:56 crc kubenswrapper[4959]: I1007 15:23:56.259000 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmrmg\" (UniqueName: \"kubernetes.io/projected/c34c5d88-8426-4e1c-a547-4688cb55e2ae-kube-api-access-nmrmg\") pod \"certified-operators-lk5mc\" (UID: \"c34c5d88-8426-4e1c-a547-4688cb55e2ae\") " pod="openshift-marketplace/certified-operators-lk5mc" Oct 07 15:23:56 crc kubenswrapper[4959]: I1007 15:23:56.384653 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lk5mc" Oct 07 15:23:56 crc kubenswrapper[4959]: I1007 15:23:56.977340 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lk5mc"] Oct 07 15:23:57 crc kubenswrapper[4959]: I1007 15:23:57.015932 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lk5mc" event={"ID":"c34c5d88-8426-4e1c-a547-4688cb55e2ae","Type":"ContainerStarted","Data":"641e17d0a56893b83712aa015f557fd8bef0728f79b772cfbca35f09158216c9"} Oct 07 15:23:58 crc kubenswrapper[4959]: I1007 15:23:58.032402 4959 generic.go:334] "Generic (PLEG): container finished" podID="c34c5d88-8426-4e1c-a547-4688cb55e2ae" containerID="6c2341a0c31a55ba74d8dbebed742d1c2b71e73c2591966ba8d8912ea903f333" exitCode=0 Oct 07 15:23:58 crc kubenswrapper[4959]: I1007 15:23:58.032683 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lk5mc" event={"ID":"c34c5d88-8426-4e1c-a547-4688cb55e2ae","Type":"ContainerDied","Data":"6c2341a0c31a55ba74d8dbebed742d1c2b71e73c2591966ba8d8912ea903f333"} Oct 07 15:23:58 crc kubenswrapper[4959]: I1007 15:23:58.035320 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 15:24:00 crc kubenswrapper[4959]: I1007 15:24:00.055369 4959 generic.go:334] "Generic (PLEG): container finished" podID="c34c5d88-8426-4e1c-a547-4688cb55e2ae" containerID="3d236dfebba48715a5f09c4791ec40d6558ee37a3d629c665e91938310f004c4" exitCode=0 Oct 07 15:24:00 crc kubenswrapper[4959]: I1007 15:24:00.055480 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lk5mc" event={"ID":"c34c5d88-8426-4e1c-a547-4688cb55e2ae","Type":"ContainerDied","Data":"3d236dfebba48715a5f09c4791ec40d6558ee37a3d629c665e91938310f004c4"} Oct 07 15:24:01 crc kubenswrapper[4959]: I1007 15:24:01.069368 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lk5mc" event={"ID":"c34c5d88-8426-4e1c-a547-4688cb55e2ae","Type":"ContainerStarted","Data":"f7d66682051bb81fb83e673dd8ae6bf558c2a704b90e801160138f05e3baeb5e"} Oct 07 15:24:01 crc kubenswrapper[4959]: I1007 15:24:01.102209 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lk5mc" podStartSLOduration=2.592991067 podStartE2EDuration="5.102171535s" podCreationTimestamp="2025-10-07 15:23:56 +0000 UTC" firstStartedPulling="2025-10-07 15:23:58.035023992 +0000 UTC m=+5920.118428307" lastFinishedPulling="2025-10-07 15:24:00.54420445 +0000 UTC m=+5922.627608775" observedRunningTime="2025-10-07 15:24:01.088828034 +0000 UTC m=+5923.172232359" watchObservedRunningTime="2025-10-07 15:24:01.102171535 +0000 UTC m=+5923.185575850" Oct 07 15:24:03 crc kubenswrapper[4959]: I1007 15:24:03.654767 4959 scope.go:117] "RemoveContainer" containerID="20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168" Oct 07 15:24:03 crc kubenswrapper[4959]: E1007 15:24:03.655632 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:24:06 crc kubenswrapper[4959]: I1007 15:24:06.387679 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lk5mc" Oct 07 15:24:06 crc kubenswrapper[4959]: I1007 15:24:06.388206 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lk5mc" Oct 07 15:24:06 crc kubenswrapper[4959]: I1007 15:24:06.437982 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lk5mc" Oct 07 15:24:07 crc kubenswrapper[4959]: I1007 15:24:07.187113 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lk5mc" Oct 07 15:24:07 crc kubenswrapper[4959]: I1007 15:24:07.247256 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lk5mc"] Oct 07 15:24:09 crc kubenswrapper[4959]: I1007 15:24:09.157172 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lk5mc" podUID="c34c5d88-8426-4e1c-a547-4688cb55e2ae" containerName="registry-server" containerID="cri-o://f7d66682051bb81fb83e673dd8ae6bf558c2a704b90e801160138f05e3baeb5e" gracePeriod=2 Oct 07 15:24:09 crc kubenswrapper[4959]: I1007 15:24:09.664968 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lk5mc" Oct 07 15:24:09 crc kubenswrapper[4959]: I1007 15:24:09.748814 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c34c5d88-8426-4e1c-a547-4688cb55e2ae-utilities\") pod \"c34c5d88-8426-4e1c-a547-4688cb55e2ae\" (UID: \"c34c5d88-8426-4e1c-a547-4688cb55e2ae\") " Oct 07 15:24:09 crc kubenswrapper[4959]: I1007 15:24:09.749147 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmrmg\" (UniqueName: \"kubernetes.io/projected/c34c5d88-8426-4e1c-a547-4688cb55e2ae-kube-api-access-nmrmg\") pod \"c34c5d88-8426-4e1c-a547-4688cb55e2ae\" (UID: \"c34c5d88-8426-4e1c-a547-4688cb55e2ae\") " Oct 07 15:24:09 crc kubenswrapper[4959]: I1007 15:24:09.750498 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c34c5d88-8426-4e1c-a547-4688cb55e2ae-catalog-content\") pod \"c34c5d88-8426-4e1c-a547-4688cb55e2ae\" (UID: \"c34c5d88-8426-4e1c-a547-4688cb55e2ae\") " Oct 07 15:24:09 crc kubenswrapper[4959]: I1007 15:24:09.750507 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c34c5d88-8426-4e1c-a547-4688cb55e2ae-utilities" (OuterVolumeSpecName: "utilities") pod "c34c5d88-8426-4e1c-a547-4688cb55e2ae" (UID: "c34c5d88-8426-4e1c-a547-4688cb55e2ae"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:24:09 crc kubenswrapper[4959]: I1007 15:24:09.751232 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c34c5d88-8426-4e1c-a547-4688cb55e2ae-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 15:24:09 crc kubenswrapper[4959]: I1007 15:24:09.758263 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c34c5d88-8426-4e1c-a547-4688cb55e2ae-kube-api-access-nmrmg" (OuterVolumeSpecName: "kube-api-access-nmrmg") pod "c34c5d88-8426-4e1c-a547-4688cb55e2ae" (UID: "c34c5d88-8426-4e1c-a547-4688cb55e2ae"). InnerVolumeSpecName "kube-api-access-nmrmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 15:24:09 crc kubenswrapper[4959]: I1007 15:24:09.853146 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmrmg\" (UniqueName: \"kubernetes.io/projected/c34c5d88-8426-4e1c-a547-4688cb55e2ae-kube-api-access-nmrmg\") on node \"crc\" DevicePath \"\"" Oct 07 15:24:10 crc kubenswrapper[4959]: I1007 15:24:10.170615 4959 generic.go:334] "Generic (PLEG): container finished" podID="c34c5d88-8426-4e1c-a547-4688cb55e2ae" containerID="f7d66682051bb81fb83e673dd8ae6bf558c2a704b90e801160138f05e3baeb5e" exitCode=0 Oct 07 15:24:10 crc kubenswrapper[4959]: I1007 15:24:10.170979 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lk5mc" event={"ID":"c34c5d88-8426-4e1c-a547-4688cb55e2ae","Type":"ContainerDied","Data":"f7d66682051bb81fb83e673dd8ae6bf558c2a704b90e801160138f05e3baeb5e"} Oct 07 15:24:10 crc kubenswrapper[4959]: I1007 15:24:10.171071 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lk5mc" event={"ID":"c34c5d88-8426-4e1c-a547-4688cb55e2ae","Type":"ContainerDied","Data":"641e17d0a56893b83712aa015f557fd8bef0728f79b772cfbca35f09158216c9"} Oct 07 15:24:10 crc kubenswrapper[4959]: I1007 15:24:10.171116 4959 scope.go:117] "RemoveContainer" containerID="f7d66682051bb81fb83e673dd8ae6bf558c2a704b90e801160138f05e3baeb5e" Oct 07 15:24:10 crc kubenswrapper[4959]: I1007 15:24:10.171256 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lk5mc" Oct 07 15:24:10 crc kubenswrapper[4959]: I1007 15:24:10.199240 4959 scope.go:117] "RemoveContainer" containerID="3d236dfebba48715a5f09c4791ec40d6558ee37a3d629c665e91938310f004c4" Oct 07 15:24:10 crc kubenswrapper[4959]: I1007 15:24:10.232221 4959 scope.go:117] "RemoveContainer" containerID="6c2341a0c31a55ba74d8dbebed742d1c2b71e73c2591966ba8d8912ea903f333" Oct 07 15:24:10 crc kubenswrapper[4959]: I1007 15:24:10.303669 4959 scope.go:117] "RemoveContainer" containerID="f7d66682051bb81fb83e673dd8ae6bf558c2a704b90e801160138f05e3baeb5e" Oct 07 15:24:10 crc kubenswrapper[4959]: E1007 15:24:10.304226 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7d66682051bb81fb83e673dd8ae6bf558c2a704b90e801160138f05e3baeb5e\": container with ID starting with f7d66682051bb81fb83e673dd8ae6bf558c2a704b90e801160138f05e3baeb5e not found: ID does not exist" containerID="f7d66682051bb81fb83e673dd8ae6bf558c2a704b90e801160138f05e3baeb5e" Oct 07 15:24:10 crc kubenswrapper[4959]: I1007 15:24:10.304279 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7d66682051bb81fb83e673dd8ae6bf558c2a704b90e801160138f05e3baeb5e"} err="failed to get container status \"f7d66682051bb81fb83e673dd8ae6bf558c2a704b90e801160138f05e3baeb5e\": rpc error: code = NotFound desc = could not find container \"f7d66682051bb81fb83e673dd8ae6bf558c2a704b90e801160138f05e3baeb5e\": container with ID starting with f7d66682051bb81fb83e673dd8ae6bf558c2a704b90e801160138f05e3baeb5e not found: ID does not exist" Oct 07 15:24:10 crc kubenswrapper[4959]: I1007 15:24:10.304308 4959 scope.go:117] "RemoveContainer" containerID="3d236dfebba48715a5f09c4791ec40d6558ee37a3d629c665e91938310f004c4" Oct 07 15:24:10 crc kubenswrapper[4959]: E1007 15:24:10.305227 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d236dfebba48715a5f09c4791ec40d6558ee37a3d629c665e91938310f004c4\": container with ID starting with 3d236dfebba48715a5f09c4791ec40d6558ee37a3d629c665e91938310f004c4 not found: ID does not exist" containerID="3d236dfebba48715a5f09c4791ec40d6558ee37a3d629c665e91938310f004c4" Oct 07 15:24:10 crc kubenswrapper[4959]: I1007 15:24:10.305281 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d236dfebba48715a5f09c4791ec40d6558ee37a3d629c665e91938310f004c4"} err="failed to get container status \"3d236dfebba48715a5f09c4791ec40d6558ee37a3d629c665e91938310f004c4\": rpc error: code = NotFound desc = could not find container \"3d236dfebba48715a5f09c4791ec40d6558ee37a3d629c665e91938310f004c4\": container with ID starting with 3d236dfebba48715a5f09c4791ec40d6558ee37a3d629c665e91938310f004c4 not found: ID does not exist" Oct 07 15:24:10 crc kubenswrapper[4959]: I1007 15:24:10.305311 4959 scope.go:117] "RemoveContainer" containerID="6c2341a0c31a55ba74d8dbebed742d1c2b71e73c2591966ba8d8912ea903f333" Oct 07 15:24:10 crc kubenswrapper[4959]: E1007 15:24:10.306045 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c2341a0c31a55ba74d8dbebed742d1c2b71e73c2591966ba8d8912ea903f333\": container with ID starting with 6c2341a0c31a55ba74d8dbebed742d1c2b71e73c2591966ba8d8912ea903f333 not found: ID does not exist" containerID="6c2341a0c31a55ba74d8dbebed742d1c2b71e73c2591966ba8d8912ea903f333" Oct 07 15:24:10 crc kubenswrapper[4959]: I1007 15:24:10.306220 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c2341a0c31a55ba74d8dbebed742d1c2b71e73c2591966ba8d8912ea903f333"} err="failed to get container status \"6c2341a0c31a55ba74d8dbebed742d1c2b71e73c2591966ba8d8912ea903f333\": rpc error: code = NotFound desc = could not find container \"6c2341a0c31a55ba74d8dbebed742d1c2b71e73c2591966ba8d8912ea903f333\": container with ID starting with 6c2341a0c31a55ba74d8dbebed742d1c2b71e73c2591966ba8d8912ea903f333 not found: ID does not exist" Oct 07 15:24:10 crc kubenswrapper[4959]: I1007 15:24:10.395290 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c34c5d88-8426-4e1c-a547-4688cb55e2ae-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c34c5d88-8426-4e1c-a547-4688cb55e2ae" (UID: "c34c5d88-8426-4e1c-a547-4688cb55e2ae"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:24:10 crc kubenswrapper[4959]: I1007 15:24:10.466637 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c34c5d88-8426-4e1c-a547-4688cb55e2ae-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 15:24:10 crc kubenswrapper[4959]: I1007 15:24:10.516436 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lk5mc"] Oct 07 15:24:10 crc kubenswrapper[4959]: I1007 15:24:10.526419 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lk5mc"] Oct 07 15:24:10 crc kubenswrapper[4959]: I1007 15:24:10.666348 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c34c5d88-8426-4e1c-a547-4688cb55e2ae" path="/var/lib/kubelet/pods/c34c5d88-8426-4e1c-a547-4688cb55e2ae/volumes" Oct 07 15:24:14 crc kubenswrapper[4959]: I1007 15:24:14.654006 4959 scope.go:117] "RemoveContainer" containerID="20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168" Oct 07 15:24:14 crc kubenswrapper[4959]: E1007 15:24:14.655174 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:24:29 crc kubenswrapper[4959]: I1007 15:24:29.653837 4959 scope.go:117] "RemoveContainer" containerID="20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168" Oct 07 15:24:29 crc kubenswrapper[4959]: E1007 15:24:29.654717 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:24:41 crc kubenswrapper[4959]: I1007 15:24:41.653971 4959 scope.go:117] "RemoveContainer" containerID="20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168" Oct 07 15:24:41 crc kubenswrapper[4959]: E1007 15:24:41.655186 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:24:56 crc kubenswrapper[4959]: I1007 15:24:56.653767 4959 scope.go:117] "RemoveContainer" containerID="20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168" Oct 07 15:24:56 crc kubenswrapper[4959]: E1007 15:24:56.654770 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:25:07 crc kubenswrapper[4959]: I1007 15:25:07.654269 4959 scope.go:117] "RemoveContainer" containerID="20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168" Oct 07 15:25:07 crc kubenswrapper[4959]: E1007 15:25:07.655018 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:25:22 crc kubenswrapper[4959]: I1007 15:25:22.653123 4959 scope.go:117] "RemoveContainer" containerID="20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168" Oct 07 15:25:22 crc kubenswrapper[4959]: E1007 15:25:22.654139 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:25:33 crc kubenswrapper[4959]: I1007 15:25:33.653632 4959 scope.go:117] "RemoveContainer" containerID="20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168" Oct 07 15:25:33 crc kubenswrapper[4959]: E1007 15:25:33.654771 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:25:45 crc kubenswrapper[4959]: I1007 15:25:45.653281 4959 scope.go:117] "RemoveContainer" containerID="20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168" Oct 07 15:25:45 crc kubenswrapper[4959]: E1007 15:25:45.654249 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:25:59 crc kubenswrapper[4959]: I1007 15:25:59.654327 4959 scope.go:117] "RemoveContainer" containerID="20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168" Oct 07 15:25:59 crc kubenswrapper[4959]: E1007 15:25:59.655399 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:26:12 crc kubenswrapper[4959]: I1007 15:26:12.654245 4959 scope.go:117] "RemoveContainer" containerID="20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168" Oct 07 15:26:12 crc kubenswrapper[4959]: E1007 15:26:12.655161 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:26:23 crc kubenswrapper[4959]: I1007 15:26:23.924520 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dbbdc"] Oct 07 15:26:23 crc kubenswrapper[4959]: E1007 15:26:23.925810 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c34c5d88-8426-4e1c-a547-4688cb55e2ae" containerName="registry-server" Oct 07 15:26:23 crc kubenswrapper[4959]: I1007 15:26:23.925827 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c34c5d88-8426-4e1c-a547-4688cb55e2ae" containerName="registry-server" Oct 07 15:26:23 crc kubenswrapper[4959]: E1007 15:26:23.925847 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c34c5d88-8426-4e1c-a547-4688cb55e2ae" containerName="extract-utilities" Oct 07 15:26:23 crc kubenswrapper[4959]: I1007 15:26:23.925854 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c34c5d88-8426-4e1c-a547-4688cb55e2ae" containerName="extract-utilities" Oct 07 15:26:23 crc kubenswrapper[4959]: E1007 15:26:23.925915 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c34c5d88-8426-4e1c-a547-4688cb55e2ae" containerName="extract-content" Oct 07 15:26:23 crc kubenswrapper[4959]: I1007 15:26:23.925922 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c34c5d88-8426-4e1c-a547-4688cb55e2ae" containerName="extract-content" Oct 07 15:26:23 crc kubenswrapper[4959]: I1007 15:26:23.926204 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="c34c5d88-8426-4e1c-a547-4688cb55e2ae" containerName="registry-server" Oct 07 15:26:23 crc kubenswrapper[4959]: I1007 15:26:23.927780 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dbbdc" Oct 07 15:26:23 crc kubenswrapper[4959]: I1007 15:26:23.952347 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dbbdc"] Oct 07 15:26:24 crc kubenswrapper[4959]: I1007 15:26:24.004493 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7d7cef6-18c1-4a69-886b-ce98ddae0858-catalog-content\") pod \"redhat-operators-dbbdc\" (UID: \"d7d7cef6-18c1-4a69-886b-ce98ddae0858\") " pod="openshift-marketplace/redhat-operators-dbbdc" Oct 07 15:26:24 crc kubenswrapper[4959]: I1007 15:26:24.004563 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7d7cef6-18c1-4a69-886b-ce98ddae0858-utilities\") pod \"redhat-operators-dbbdc\" (UID: \"d7d7cef6-18c1-4a69-886b-ce98ddae0858\") " pod="openshift-marketplace/redhat-operators-dbbdc" Oct 07 15:26:24 crc kubenswrapper[4959]: I1007 15:26:24.004594 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xhk2\" (UniqueName: \"kubernetes.io/projected/d7d7cef6-18c1-4a69-886b-ce98ddae0858-kube-api-access-5xhk2\") pod \"redhat-operators-dbbdc\" (UID: \"d7d7cef6-18c1-4a69-886b-ce98ddae0858\") " pod="openshift-marketplace/redhat-operators-dbbdc" Oct 07 15:26:24 crc kubenswrapper[4959]: I1007 15:26:24.106456 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7d7cef6-18c1-4a69-886b-ce98ddae0858-utilities\") pod \"redhat-operators-dbbdc\" (UID: \"d7d7cef6-18c1-4a69-886b-ce98ddae0858\") " pod="openshift-marketplace/redhat-operators-dbbdc" Oct 07 15:26:24 crc kubenswrapper[4959]: I1007 15:26:24.106524 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xhk2\" (UniqueName: \"kubernetes.io/projected/d7d7cef6-18c1-4a69-886b-ce98ddae0858-kube-api-access-5xhk2\") pod \"redhat-operators-dbbdc\" (UID: \"d7d7cef6-18c1-4a69-886b-ce98ddae0858\") " pod="openshift-marketplace/redhat-operators-dbbdc" Oct 07 15:26:24 crc kubenswrapper[4959]: I1007 15:26:24.106739 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7d7cef6-18c1-4a69-886b-ce98ddae0858-catalog-content\") pod \"redhat-operators-dbbdc\" (UID: \"d7d7cef6-18c1-4a69-886b-ce98ddae0858\") " pod="openshift-marketplace/redhat-operators-dbbdc" Oct 07 15:26:24 crc kubenswrapper[4959]: I1007 15:26:24.107003 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7d7cef6-18c1-4a69-886b-ce98ddae0858-utilities\") pod \"redhat-operators-dbbdc\" (UID: \"d7d7cef6-18c1-4a69-886b-ce98ddae0858\") " pod="openshift-marketplace/redhat-operators-dbbdc" Oct 07 15:26:24 crc kubenswrapper[4959]: I1007 15:26:24.107949 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7d7cef6-18c1-4a69-886b-ce98ddae0858-catalog-content\") pod \"redhat-operators-dbbdc\" (UID: \"d7d7cef6-18c1-4a69-886b-ce98ddae0858\") " pod="openshift-marketplace/redhat-operators-dbbdc" Oct 07 15:26:24 crc kubenswrapper[4959]: I1007 15:26:24.131129 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xhk2\" (UniqueName: \"kubernetes.io/projected/d7d7cef6-18c1-4a69-886b-ce98ddae0858-kube-api-access-5xhk2\") pod \"redhat-operators-dbbdc\" (UID: \"d7d7cef6-18c1-4a69-886b-ce98ddae0858\") " pod="openshift-marketplace/redhat-operators-dbbdc" Oct 07 15:26:24 crc kubenswrapper[4959]: I1007 15:26:24.249973 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dbbdc" Oct 07 15:26:24 crc kubenswrapper[4959]: I1007 15:26:24.756273 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dbbdc"] Oct 07 15:26:25 crc kubenswrapper[4959]: I1007 15:26:25.442875 4959 generic.go:334] "Generic (PLEG): container finished" podID="d7d7cef6-18c1-4a69-886b-ce98ddae0858" containerID="63b17a5a823a6585983f35fa3301b55ec7975fa85c661f84b5d9b04558b00180" exitCode=0 Oct 07 15:26:25 crc kubenswrapper[4959]: I1007 15:26:25.443292 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dbbdc" event={"ID":"d7d7cef6-18c1-4a69-886b-ce98ddae0858","Type":"ContainerDied","Data":"63b17a5a823a6585983f35fa3301b55ec7975fa85c661f84b5d9b04558b00180"} Oct 07 15:26:25 crc kubenswrapper[4959]: I1007 15:26:25.443323 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dbbdc" event={"ID":"d7d7cef6-18c1-4a69-886b-ce98ddae0858","Type":"ContainerStarted","Data":"b70d5fd973cf4969f53db7d7137f61df311c453381e670adabf23b70ce95d117"} Oct 07 15:26:27 crc kubenswrapper[4959]: I1007 15:26:27.465795 4959 generic.go:334] "Generic (PLEG): container finished" podID="d7d7cef6-18c1-4a69-886b-ce98ddae0858" containerID="e9ee17b8f1b7cba31b95600fb4ec4a538ec5ee8687add776541b0549f304e06b" exitCode=0 Oct 07 15:26:27 crc kubenswrapper[4959]: I1007 15:26:27.465914 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dbbdc" event={"ID":"d7d7cef6-18c1-4a69-886b-ce98ddae0858","Type":"ContainerDied","Data":"e9ee17b8f1b7cba31b95600fb4ec4a538ec5ee8687add776541b0549f304e06b"} Oct 07 15:26:27 crc kubenswrapper[4959]: I1007 15:26:27.653660 4959 scope.go:117] "RemoveContainer" containerID="20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168" Oct 07 15:26:28 crc kubenswrapper[4959]: I1007 15:26:28.483316 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerStarted","Data":"3b696f2313816349c23c9e18632aa1ab844c286a6de0e092da7dc16410d9a66e"} Oct 07 15:26:28 crc kubenswrapper[4959]: I1007 15:26:28.488724 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dbbdc" event={"ID":"d7d7cef6-18c1-4a69-886b-ce98ddae0858","Type":"ContainerStarted","Data":"941b9bfcd0677b007273da1eb7fc0a46fb9cc409655a467e86e8375251c9efd0"} Oct 07 15:26:28 crc kubenswrapper[4959]: I1007 15:26:28.531825 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dbbdc" podStartSLOduration=3.070536114 podStartE2EDuration="5.531801426s" podCreationTimestamp="2025-10-07 15:26:23 +0000 UTC" firstStartedPulling="2025-10-07 15:26:25.444981999 +0000 UTC m=+6067.528386324" lastFinishedPulling="2025-10-07 15:26:27.906247311 +0000 UTC m=+6069.989651636" observedRunningTime="2025-10-07 15:26:28.520713883 +0000 UTC m=+6070.604118228" watchObservedRunningTime="2025-10-07 15:26:28.531801426 +0000 UTC m=+6070.615205751" Oct 07 15:26:34 crc kubenswrapper[4959]: I1007 15:26:34.250815 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dbbdc" Oct 07 15:26:34 crc kubenswrapper[4959]: I1007 15:26:34.251652 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dbbdc" Oct 07 15:26:34 crc kubenswrapper[4959]: I1007 15:26:34.296228 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dbbdc" Oct 07 15:26:34 crc kubenswrapper[4959]: I1007 15:26:34.591546 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dbbdc" Oct 07 15:26:34 crc kubenswrapper[4959]: I1007 15:26:34.648366 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dbbdc"] Oct 07 15:26:36 crc kubenswrapper[4959]: I1007 15:26:36.557943 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dbbdc" podUID="d7d7cef6-18c1-4a69-886b-ce98ddae0858" containerName="registry-server" containerID="cri-o://941b9bfcd0677b007273da1eb7fc0a46fb9cc409655a467e86e8375251c9efd0" gracePeriod=2 Oct 07 15:26:37 crc kubenswrapper[4959]: I1007 15:26:37.078074 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dbbdc" Oct 07 15:26:37 crc kubenswrapper[4959]: I1007 15:26:37.177514 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xhk2\" (UniqueName: \"kubernetes.io/projected/d7d7cef6-18c1-4a69-886b-ce98ddae0858-kube-api-access-5xhk2\") pod \"d7d7cef6-18c1-4a69-886b-ce98ddae0858\" (UID: \"d7d7cef6-18c1-4a69-886b-ce98ddae0858\") " Oct 07 15:26:37 crc kubenswrapper[4959]: I1007 15:26:37.177689 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7d7cef6-18c1-4a69-886b-ce98ddae0858-utilities\") pod \"d7d7cef6-18c1-4a69-886b-ce98ddae0858\" (UID: \"d7d7cef6-18c1-4a69-886b-ce98ddae0858\") " Oct 07 15:26:37 crc kubenswrapper[4959]: I1007 15:26:37.177784 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7d7cef6-18c1-4a69-886b-ce98ddae0858-catalog-content\") pod \"d7d7cef6-18c1-4a69-886b-ce98ddae0858\" (UID: \"d7d7cef6-18c1-4a69-886b-ce98ddae0858\") " Oct 07 15:26:37 crc kubenswrapper[4959]: I1007 15:26:37.178734 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7d7cef6-18c1-4a69-886b-ce98ddae0858-utilities" (OuterVolumeSpecName: "utilities") pod "d7d7cef6-18c1-4a69-886b-ce98ddae0858" (UID: "d7d7cef6-18c1-4a69-886b-ce98ddae0858"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:26:37 crc kubenswrapper[4959]: I1007 15:26:37.183507 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7d7cef6-18c1-4a69-886b-ce98ddae0858-kube-api-access-5xhk2" (OuterVolumeSpecName: "kube-api-access-5xhk2") pod "d7d7cef6-18c1-4a69-886b-ce98ddae0858" (UID: "d7d7cef6-18c1-4a69-886b-ce98ddae0858"). InnerVolumeSpecName "kube-api-access-5xhk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 15:26:37 crc kubenswrapper[4959]: I1007 15:26:37.280417 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xhk2\" (UniqueName: \"kubernetes.io/projected/d7d7cef6-18c1-4a69-886b-ce98ddae0858-kube-api-access-5xhk2\") on node \"crc\" DevicePath \"\"" Oct 07 15:26:37 crc kubenswrapper[4959]: I1007 15:26:37.280463 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7d7cef6-18c1-4a69-886b-ce98ddae0858-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 15:26:37 crc kubenswrapper[4959]: I1007 15:26:37.569712 4959 generic.go:334] "Generic (PLEG): container finished" podID="d7d7cef6-18c1-4a69-886b-ce98ddae0858" containerID="941b9bfcd0677b007273da1eb7fc0a46fb9cc409655a467e86e8375251c9efd0" exitCode=0 Oct 07 15:26:37 crc kubenswrapper[4959]: I1007 15:26:37.569801 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dbbdc" event={"ID":"d7d7cef6-18c1-4a69-886b-ce98ddae0858","Type":"ContainerDied","Data":"941b9bfcd0677b007273da1eb7fc0a46fb9cc409655a467e86e8375251c9efd0"} Oct 07 15:26:37 crc kubenswrapper[4959]: I1007 15:26:37.570159 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dbbdc" event={"ID":"d7d7cef6-18c1-4a69-886b-ce98ddae0858","Type":"ContainerDied","Data":"b70d5fd973cf4969f53db7d7137f61df311c453381e670adabf23b70ce95d117"} Oct 07 15:26:37 crc kubenswrapper[4959]: I1007 15:26:37.570186 4959 scope.go:117] "RemoveContainer" containerID="941b9bfcd0677b007273da1eb7fc0a46fb9cc409655a467e86e8375251c9efd0" Oct 07 15:26:37 crc kubenswrapper[4959]: I1007 15:26:37.569880 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dbbdc" Oct 07 15:26:37 crc kubenswrapper[4959]: I1007 15:26:37.600283 4959 scope.go:117] "RemoveContainer" containerID="e9ee17b8f1b7cba31b95600fb4ec4a538ec5ee8687add776541b0549f304e06b" Oct 07 15:26:37 crc kubenswrapper[4959]: I1007 15:26:37.624438 4959 scope.go:117] "RemoveContainer" containerID="63b17a5a823a6585983f35fa3301b55ec7975fa85c661f84b5d9b04558b00180" Oct 07 15:26:37 crc kubenswrapper[4959]: I1007 15:26:37.678837 4959 scope.go:117] "RemoveContainer" containerID="941b9bfcd0677b007273da1eb7fc0a46fb9cc409655a467e86e8375251c9efd0" Oct 07 15:26:37 crc kubenswrapper[4959]: E1007 15:26:37.679633 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"941b9bfcd0677b007273da1eb7fc0a46fb9cc409655a467e86e8375251c9efd0\": container with ID starting with 941b9bfcd0677b007273da1eb7fc0a46fb9cc409655a467e86e8375251c9efd0 not found: ID does not exist" containerID="941b9bfcd0677b007273da1eb7fc0a46fb9cc409655a467e86e8375251c9efd0" Oct 07 15:26:37 crc kubenswrapper[4959]: I1007 15:26:37.679678 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"941b9bfcd0677b007273da1eb7fc0a46fb9cc409655a467e86e8375251c9efd0"} err="failed to get container status \"941b9bfcd0677b007273da1eb7fc0a46fb9cc409655a467e86e8375251c9efd0\": rpc error: code = NotFound desc = could not find container \"941b9bfcd0677b007273da1eb7fc0a46fb9cc409655a467e86e8375251c9efd0\": container with ID starting with 941b9bfcd0677b007273da1eb7fc0a46fb9cc409655a467e86e8375251c9efd0 not found: ID does not exist" Oct 07 15:26:37 crc kubenswrapper[4959]: I1007 15:26:37.679709 4959 scope.go:117] "RemoveContainer" containerID="e9ee17b8f1b7cba31b95600fb4ec4a538ec5ee8687add776541b0549f304e06b" Oct 07 15:26:37 crc kubenswrapper[4959]: E1007 15:26:37.680143 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9ee17b8f1b7cba31b95600fb4ec4a538ec5ee8687add776541b0549f304e06b\": container with ID starting with e9ee17b8f1b7cba31b95600fb4ec4a538ec5ee8687add776541b0549f304e06b not found: ID does not exist" containerID="e9ee17b8f1b7cba31b95600fb4ec4a538ec5ee8687add776541b0549f304e06b" Oct 07 15:26:37 crc kubenswrapper[4959]: I1007 15:26:37.680166 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9ee17b8f1b7cba31b95600fb4ec4a538ec5ee8687add776541b0549f304e06b"} err="failed to get container status \"e9ee17b8f1b7cba31b95600fb4ec4a538ec5ee8687add776541b0549f304e06b\": rpc error: code = NotFound desc = could not find container \"e9ee17b8f1b7cba31b95600fb4ec4a538ec5ee8687add776541b0549f304e06b\": container with ID starting with e9ee17b8f1b7cba31b95600fb4ec4a538ec5ee8687add776541b0549f304e06b not found: ID does not exist" Oct 07 15:26:37 crc kubenswrapper[4959]: I1007 15:26:37.680181 4959 scope.go:117] "RemoveContainer" containerID="63b17a5a823a6585983f35fa3301b55ec7975fa85c661f84b5d9b04558b00180" Oct 07 15:26:37 crc kubenswrapper[4959]: E1007 15:26:37.680563 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63b17a5a823a6585983f35fa3301b55ec7975fa85c661f84b5d9b04558b00180\": container with ID starting with 63b17a5a823a6585983f35fa3301b55ec7975fa85c661f84b5d9b04558b00180 not found: ID does not exist" containerID="63b17a5a823a6585983f35fa3301b55ec7975fa85c661f84b5d9b04558b00180" Oct 07 15:26:37 crc kubenswrapper[4959]: I1007 15:26:37.680618 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63b17a5a823a6585983f35fa3301b55ec7975fa85c661f84b5d9b04558b00180"} err="failed to get container status \"63b17a5a823a6585983f35fa3301b55ec7975fa85c661f84b5d9b04558b00180\": rpc error: code = NotFound desc = could not find container \"63b17a5a823a6585983f35fa3301b55ec7975fa85c661f84b5d9b04558b00180\": container with ID starting with 63b17a5a823a6585983f35fa3301b55ec7975fa85c661f84b5d9b04558b00180 not found: ID does not exist" Oct 07 15:26:38 crc kubenswrapper[4959]: I1007 15:26:38.705212 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7d7cef6-18c1-4a69-886b-ce98ddae0858-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d7d7cef6-18c1-4a69-886b-ce98ddae0858" (UID: "d7d7cef6-18c1-4a69-886b-ce98ddae0858"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:26:38 crc kubenswrapper[4959]: I1007 15:26:38.711424 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7d7cef6-18c1-4a69-886b-ce98ddae0858-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 15:26:38 crc kubenswrapper[4959]: I1007 15:26:38.805922 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dbbdc"] Oct 07 15:26:38 crc kubenswrapper[4959]: I1007 15:26:38.814259 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dbbdc"] Oct 07 15:26:40 crc kubenswrapper[4959]: I1007 15:26:40.664814 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7d7cef6-18c1-4a69-886b-ce98ddae0858" path="/var/lib/kubelet/pods/d7d7cef6-18c1-4a69-886b-ce98ddae0858/volumes" Oct 07 15:26:43 crc kubenswrapper[4959]: I1007 15:26:43.632248 4959 generic.go:334] "Generic (PLEG): container finished" podID="bc180a96-fac6-4b97-97bf-11b7c7ceff8a" containerID="83923ad3156f01c3fd190fb9bddab0839d9f7a15acf5197b50ba35bd871a4fe5" exitCode=0 Oct 07 15:26:43 crc kubenswrapper[4959]: I1007 15:26:43.632301 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-test" event={"ID":"bc180a96-fac6-4b97-97bf-11b7c7ceff8a","Type":"ContainerDied","Data":"83923ad3156f01c3fd190fb9bddab0839d9f7a15acf5197b50ba35bd871a4fe5"} Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.122346 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.159672 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-openstack-config-secret\") pod \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.159738 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-test-operator-ephemeral-workdir\") pod \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.159798 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-config-data\") pod \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.159845 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-ca-certs\") pod \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.159915 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-openstack-config\") pod \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.159995 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n89v8\" (UniqueName: \"kubernetes.io/projected/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-kube-api-access-n89v8\") pod \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.160034 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-test-operator-ephemeral-temporary\") pod \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.160069 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-ssh-key\") pod \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.160146 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.160185 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-ceph\") pod \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\" (UID: \"bc180a96-fac6-4b97-97bf-11b7c7ceff8a\") " Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.160975 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "bc180a96-fac6-4b97-97bf-11b7c7ceff8a" (UID: "bc180a96-fac6-4b97-97bf-11b7c7ceff8a"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.161358 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-config-data" (OuterVolumeSpecName: "config-data") pod "bc180a96-fac6-4b97-97bf-11b7c7ceff8a" (UID: "bc180a96-fac6-4b97-97bf-11b7c7ceff8a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.168238 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-ceph" (OuterVolumeSpecName: "ceph") pod "bc180a96-fac6-4b97-97bf-11b7c7ceff8a" (UID: "bc180a96-fac6-4b97-97bf-11b7c7ceff8a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.169086 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-kube-api-access-n89v8" (OuterVolumeSpecName: "kube-api-access-n89v8") pod "bc180a96-fac6-4b97-97bf-11b7c7ceff8a" (UID: "bc180a96-fac6-4b97-97bf-11b7c7ceff8a"). InnerVolumeSpecName "kube-api-access-n89v8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.169796 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "test-operator-logs") pod "bc180a96-fac6-4b97-97bf-11b7c7ceff8a" (UID: "bc180a96-fac6-4b97-97bf-11b7c7ceff8a"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.176579 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "bc180a96-fac6-4b97-97bf-11b7c7ceff8a" (UID: "bc180a96-fac6-4b97-97bf-11b7c7ceff8a"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.195848 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "bc180a96-fac6-4b97-97bf-11b7c7ceff8a" (UID: "bc180a96-fac6-4b97-97bf-11b7c7ceff8a"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.198165 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "bc180a96-fac6-4b97-97bf-11b7c7ceff8a" (UID: "bc180a96-fac6-4b97-97bf-11b7c7ceff8a"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.206504 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bc180a96-fac6-4b97-97bf-11b7c7ceff8a" (UID: "bc180a96-fac6-4b97-97bf-11b7c7ceff8a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.224337 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "bc180a96-fac6-4b97-97bf-11b7c7ceff8a" (UID: "bc180a96-fac6-4b97-97bf-11b7c7ceff8a"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.263018 4959 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.263283 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n89v8\" (UniqueName: \"kubernetes.io/projected/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-kube-api-access-n89v8\") on node \"crc\" DevicePath \"\"" Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.263392 4959 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.263464 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.263559 4959 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.263621 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-ceph\") on node \"crc\" DevicePath \"\"" Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.263686 4959 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.263745 4959 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.263806 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-config-data\") on node \"crc\" DevicePath \"\"" Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.263864 4959 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/bc180a96-fac6-4b97-97bf-11b7c7ceff8a-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.286622 4959 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.366401 4959 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.650845 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-test" event={"ID":"bc180a96-fac6-4b97-97bf-11b7c7ceff8a","Type":"ContainerDied","Data":"e23f8218120916894a8f6c03f741cc196294af851ffa4762217348e4c312191e"} Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.650879 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-test" Oct 07 15:26:45 crc kubenswrapper[4959]: I1007 15:26:45.650888 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e23f8218120916894a8f6c03f741cc196294af851ffa4762217348e4c312191e" Oct 07 15:26:55 crc kubenswrapper[4959]: I1007 15:26:55.110815 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 07 15:26:55 crc kubenswrapper[4959]: E1007 15:26:55.112301 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7d7cef6-18c1-4a69-886b-ce98ddae0858" containerName="extract-utilities" Oct 07 15:26:55 crc kubenswrapper[4959]: I1007 15:26:55.112320 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7d7cef6-18c1-4a69-886b-ce98ddae0858" containerName="extract-utilities" Oct 07 15:26:55 crc kubenswrapper[4959]: E1007 15:26:55.112354 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7d7cef6-18c1-4a69-886b-ce98ddae0858" containerName="registry-server" Oct 07 15:26:55 crc kubenswrapper[4959]: I1007 15:26:55.112360 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7d7cef6-18c1-4a69-886b-ce98ddae0858" containerName="registry-server" Oct 07 15:26:55 crc kubenswrapper[4959]: E1007 15:26:55.112382 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc180a96-fac6-4b97-97bf-11b7c7ceff8a" containerName="tempest-tests-tempest-tests-runner" Oct 07 15:26:55 crc kubenswrapper[4959]: I1007 15:26:55.112392 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc180a96-fac6-4b97-97bf-11b7c7ceff8a" containerName="tempest-tests-tempest-tests-runner" Oct 07 15:26:55 crc kubenswrapper[4959]: E1007 15:26:55.112410 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7d7cef6-18c1-4a69-886b-ce98ddae0858" containerName="extract-content" Oct 07 15:26:55 crc kubenswrapper[4959]: I1007 15:26:55.112419 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7d7cef6-18c1-4a69-886b-ce98ddae0858" containerName="extract-content" Oct 07 15:26:55 crc kubenswrapper[4959]: I1007 15:26:55.112639 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc180a96-fac6-4b97-97bf-11b7c7ceff8a" containerName="tempest-tests-tempest-tests-runner" Oct 07 15:26:55 crc kubenswrapper[4959]: I1007 15:26:55.112663 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7d7cef6-18c1-4a69-886b-ce98ddae0858" containerName="registry-server" Oct 07 15:26:55 crc kubenswrapper[4959]: I1007 15:26:55.113627 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 07 15:26:55 crc kubenswrapper[4959]: I1007 15:26:55.115983 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-9g5qz" Oct 07 15:26:55 crc kubenswrapper[4959]: I1007 15:26:55.124489 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 07 15:26:55 crc kubenswrapper[4959]: I1007 15:26:55.281338 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pc55p\" (UniqueName: \"kubernetes.io/projected/5eb07cef-84bc-4028-afd3-b88f5ee8ef75-kube-api-access-pc55p\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5eb07cef-84bc-4028-afd3-b88f5ee8ef75\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 07 15:26:55 crc kubenswrapper[4959]: I1007 15:26:55.281429 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5eb07cef-84bc-4028-afd3-b88f5ee8ef75\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 07 15:26:55 crc kubenswrapper[4959]: I1007 15:26:55.383271 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5eb07cef-84bc-4028-afd3-b88f5ee8ef75\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 07 15:26:55 crc kubenswrapper[4959]: I1007 15:26:55.383494 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pc55p\" (UniqueName: \"kubernetes.io/projected/5eb07cef-84bc-4028-afd3-b88f5ee8ef75-kube-api-access-pc55p\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5eb07cef-84bc-4028-afd3-b88f5ee8ef75\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 07 15:26:55 crc kubenswrapper[4959]: I1007 15:26:55.384034 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5eb07cef-84bc-4028-afd3-b88f5ee8ef75\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 07 15:26:55 crc kubenswrapper[4959]: I1007 15:26:55.409276 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pc55p\" (UniqueName: \"kubernetes.io/projected/5eb07cef-84bc-4028-afd3-b88f5ee8ef75-kube-api-access-pc55p\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5eb07cef-84bc-4028-afd3-b88f5ee8ef75\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 07 15:26:55 crc kubenswrapper[4959]: I1007 15:26:55.411639 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5eb07cef-84bc-4028-afd3-b88f5ee8ef75\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 07 15:26:55 crc kubenswrapper[4959]: I1007 15:26:55.438923 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 07 15:26:55 crc kubenswrapper[4959]: I1007 15:26:55.911342 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 07 15:26:56 crc kubenswrapper[4959]: I1007 15:26:56.780770 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"5eb07cef-84bc-4028-afd3-b88f5ee8ef75","Type":"ContainerStarted","Data":"41aa378cc074a072733d237f4af6b32dac850f5299b826c12ef6a7cc5eae1abc"} Oct 07 15:26:57 crc kubenswrapper[4959]: I1007 15:26:57.791511 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"5eb07cef-84bc-4028-afd3-b88f5ee8ef75","Type":"ContainerStarted","Data":"52ad0a4960954eb1d04f8f209aa60cfa67e78a81181aeabb351870c1edc6a6bf"} Oct 07 15:26:57 crc kubenswrapper[4959]: I1007 15:26:57.805393 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.967597392 podStartE2EDuration="2.80537218s" podCreationTimestamp="2025-10-07 15:26:55 +0000 UTC" firstStartedPulling="2025-10-07 15:26:55.914524135 +0000 UTC m=+6097.997928460" lastFinishedPulling="2025-10-07 15:26:56.752298923 +0000 UTC m=+6098.835703248" observedRunningTime="2025-10-07 15:26:57.804348034 +0000 UTC m=+6099.887752359" watchObservedRunningTime="2025-10-07 15:26:57.80537218 +0000 UTC m=+6099.888776505" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.429037 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tobiko-tests-tobiko-s00-podified-functional"] Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.431285 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.438924 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tobiko-tests-tobiko-s00-podified-functional"] Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.489134 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tobiko-tests-tobikotobiko-private-key" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.489176 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"tobiko-secret" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.489195 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tobiko-tests-tobikotobiko-config" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.489293 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"test-operator-clouds-config" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.489424 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tobiko-tests-tobikotobiko-public-key" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.587570 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a646a7f2-705e-4278-9950-1a20b2db0149-openstack-config-secret\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.587701 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a646a7f2-705e-4278-9950-1a20b2db0149-ca-certs\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.587853 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a646a7f2-705e-4278-9950-1a20b2db0149-ceph\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.587934 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.588028 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8kt5\" (UniqueName: \"kubernetes.io/projected/a646a7f2-705e-4278-9950-1a20b2db0149-kube-api-access-f8kt5\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.588068 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a646a7f2-705e-4278-9950-1a20b2db0149-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.588131 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/a646a7f2-705e-4278-9950-1a20b2db0149-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.588373 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/a646a7f2-705e-4278-9950-1a20b2db0149-tobiko-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.588438 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/a646a7f2-705e-4278-9950-1a20b2db0149-kubeconfig\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.588490 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a646a7f2-705e-4278-9950-1a20b2db0149-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.588534 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/a646a7f2-705e-4278-9950-1a20b2db0149-tobiko-public-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.588559 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/a646a7f2-705e-4278-9950-1a20b2db0149-tobiko-private-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.690881 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/a646a7f2-705e-4278-9950-1a20b2db0149-tobiko-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.690959 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/a646a7f2-705e-4278-9950-1a20b2db0149-kubeconfig\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.691006 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a646a7f2-705e-4278-9950-1a20b2db0149-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.691093 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/a646a7f2-705e-4278-9950-1a20b2db0149-tobiko-public-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.691151 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/a646a7f2-705e-4278-9950-1a20b2db0149-tobiko-private-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.691290 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a646a7f2-705e-4278-9950-1a20b2db0149-openstack-config-secret\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.691320 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a646a7f2-705e-4278-9950-1a20b2db0149-ca-certs\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.691376 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a646a7f2-705e-4278-9950-1a20b2db0149-ceph\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.691417 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.691457 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8kt5\" (UniqueName: \"kubernetes.io/projected/a646a7f2-705e-4278-9950-1a20b2db0149-kube-api-access-f8kt5\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.691496 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a646a7f2-705e-4278-9950-1a20b2db0149-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.691825 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a646a7f2-705e-4278-9950-1a20b2db0149-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.692164 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a646a7f2-705e-4278-9950-1a20b2db0149-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.692178 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/a646a7f2-705e-4278-9950-1a20b2db0149-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.692173 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.692597 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/a646a7f2-705e-4278-9950-1a20b2db0149-tobiko-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.692640 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/a646a7f2-705e-4278-9950-1a20b2db0149-tobiko-public-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.692648 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/a646a7f2-705e-4278-9950-1a20b2db0149-tobiko-private-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.693332 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/a646a7f2-705e-4278-9950-1a20b2db0149-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.699414 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/a646a7f2-705e-4278-9950-1a20b2db0149-kubeconfig\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.699489 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a646a7f2-705e-4278-9950-1a20b2db0149-ca-certs\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.699671 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a646a7f2-705e-4278-9950-1a20b2db0149-openstack-config-secret\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.703533 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a646a7f2-705e-4278-9950-1a20b2db0149-ceph\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.713344 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8kt5\" (UniqueName: \"kubernetes.io/projected/a646a7f2-705e-4278-9950-1a20b2db0149-kube-api-access-f8kt5\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.731534 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:20 crc kubenswrapper[4959]: I1007 15:27:20.815172 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:27:21 crc kubenswrapper[4959]: I1007 15:27:21.351826 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tobiko-tests-tobiko-s00-podified-functional"] Oct 07 15:27:22 crc kubenswrapper[4959]: I1007 15:27:22.018429 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" event={"ID":"a646a7f2-705e-4278-9950-1a20b2db0149","Type":"ContainerStarted","Data":"5c76f28c5507de82b2666bf23c4cffaf6c42767cd17fa9e5b87bff4f143dd524"} Oct 07 15:27:38 crc kubenswrapper[4959]: I1007 15:27:38.163730 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" event={"ID":"a646a7f2-705e-4278-9950-1a20b2db0149","Type":"ContainerStarted","Data":"e8d188e7a4882290cbeb5076f525e129320a29fe0604137f84342cb01782be2b"} Oct 07 15:27:38 crc kubenswrapper[4959]: I1007 15:27:38.190786 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" podStartSLOduration=3.9325646389999998 podStartE2EDuration="19.190760594s" podCreationTimestamp="2025-10-07 15:27:19 +0000 UTC" firstStartedPulling="2025-10-07 15:27:21.367091928 +0000 UTC m=+6123.450496253" lastFinishedPulling="2025-10-07 15:27:36.625287883 +0000 UTC m=+6138.708692208" observedRunningTime="2025-10-07 15:27:38.183832907 +0000 UTC m=+6140.267237242" watchObservedRunningTime="2025-10-07 15:27:38.190760594 +0000 UTC m=+6140.274164919" Oct 07 15:27:40 crc kubenswrapper[4959]: I1007 15:27:40.232804 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-76rjc"] Oct 07 15:27:40 crc kubenswrapper[4959]: I1007 15:27:40.235155 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-76rjc" Oct 07 15:27:40 crc kubenswrapper[4959]: I1007 15:27:40.262811 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-76rjc"] Oct 07 15:27:40 crc kubenswrapper[4959]: I1007 15:27:40.339280 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cc7a017-bf74-4adb-8333-b504efcae212-catalog-content\") pod \"community-operators-76rjc\" (UID: \"8cc7a017-bf74-4adb-8333-b504efcae212\") " pod="openshift-marketplace/community-operators-76rjc" Oct 07 15:27:40 crc kubenswrapper[4959]: I1007 15:27:40.339383 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cc7a017-bf74-4adb-8333-b504efcae212-utilities\") pod \"community-operators-76rjc\" (UID: \"8cc7a017-bf74-4adb-8333-b504efcae212\") " pod="openshift-marketplace/community-operators-76rjc" Oct 07 15:27:40 crc kubenswrapper[4959]: I1007 15:27:40.339756 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlvt9\" (UniqueName: \"kubernetes.io/projected/8cc7a017-bf74-4adb-8333-b504efcae212-kube-api-access-tlvt9\") pod \"community-operators-76rjc\" (UID: \"8cc7a017-bf74-4adb-8333-b504efcae212\") " pod="openshift-marketplace/community-operators-76rjc" Oct 07 15:27:40 crc kubenswrapper[4959]: I1007 15:27:40.441667 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlvt9\" (UniqueName: \"kubernetes.io/projected/8cc7a017-bf74-4adb-8333-b504efcae212-kube-api-access-tlvt9\") pod \"community-operators-76rjc\" (UID: \"8cc7a017-bf74-4adb-8333-b504efcae212\") " pod="openshift-marketplace/community-operators-76rjc" Oct 07 15:27:40 crc kubenswrapper[4959]: I1007 15:27:40.441795 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cc7a017-bf74-4adb-8333-b504efcae212-catalog-content\") pod \"community-operators-76rjc\" (UID: \"8cc7a017-bf74-4adb-8333-b504efcae212\") " pod="openshift-marketplace/community-operators-76rjc" Oct 07 15:27:40 crc kubenswrapper[4959]: I1007 15:27:40.441871 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cc7a017-bf74-4adb-8333-b504efcae212-utilities\") pod \"community-operators-76rjc\" (UID: \"8cc7a017-bf74-4adb-8333-b504efcae212\") " pod="openshift-marketplace/community-operators-76rjc" Oct 07 15:27:40 crc kubenswrapper[4959]: I1007 15:27:40.442788 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cc7a017-bf74-4adb-8333-b504efcae212-utilities\") pod \"community-operators-76rjc\" (UID: \"8cc7a017-bf74-4adb-8333-b504efcae212\") " pod="openshift-marketplace/community-operators-76rjc" Oct 07 15:27:40 crc kubenswrapper[4959]: I1007 15:27:40.442834 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cc7a017-bf74-4adb-8333-b504efcae212-catalog-content\") pod \"community-operators-76rjc\" (UID: \"8cc7a017-bf74-4adb-8333-b504efcae212\") " pod="openshift-marketplace/community-operators-76rjc" Oct 07 15:27:40 crc kubenswrapper[4959]: I1007 15:27:40.465399 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlvt9\" (UniqueName: \"kubernetes.io/projected/8cc7a017-bf74-4adb-8333-b504efcae212-kube-api-access-tlvt9\") pod \"community-operators-76rjc\" (UID: \"8cc7a017-bf74-4adb-8333-b504efcae212\") " pod="openshift-marketplace/community-operators-76rjc" Oct 07 15:27:40 crc kubenswrapper[4959]: I1007 15:27:40.560810 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-76rjc" Oct 07 15:27:41 crc kubenswrapper[4959]: W1007 15:27:41.138876 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8cc7a017_bf74_4adb_8333_b504efcae212.slice/crio-e3127644ff187993aa02079b4096d463d52dfb7fdc96593e7e69ec11578e7983 WatchSource:0}: Error finding container e3127644ff187993aa02079b4096d463d52dfb7fdc96593e7e69ec11578e7983: Status 404 returned error can't find the container with id e3127644ff187993aa02079b4096d463d52dfb7fdc96593e7e69ec11578e7983 Oct 07 15:27:41 crc kubenswrapper[4959]: I1007 15:27:41.140560 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-76rjc"] Oct 07 15:27:41 crc kubenswrapper[4959]: I1007 15:27:41.192478 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76rjc" event={"ID":"8cc7a017-bf74-4adb-8333-b504efcae212","Type":"ContainerStarted","Data":"e3127644ff187993aa02079b4096d463d52dfb7fdc96593e7e69ec11578e7983"} Oct 07 15:27:43 crc kubenswrapper[4959]: I1007 15:27:43.211593 4959 generic.go:334] "Generic (PLEG): container finished" podID="8cc7a017-bf74-4adb-8333-b504efcae212" containerID="b870cf58fcdfb6ede6e7e75c675720cbe7dbe53f988f7988654b3322bfd990c5" exitCode=0 Oct 07 15:27:43 crc kubenswrapper[4959]: I1007 15:27:43.211685 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76rjc" event={"ID":"8cc7a017-bf74-4adb-8333-b504efcae212","Type":"ContainerDied","Data":"b870cf58fcdfb6ede6e7e75c675720cbe7dbe53f988f7988654b3322bfd990c5"} Oct 07 15:27:45 crc kubenswrapper[4959]: I1007 15:27:45.231683 4959 generic.go:334] "Generic (PLEG): container finished" podID="8cc7a017-bf74-4adb-8333-b504efcae212" containerID="eaa5453981b644977fb5bc16093c6e3c139279399ad0f955882da4e798700426" exitCode=0 Oct 07 15:27:45 crc kubenswrapper[4959]: I1007 15:27:45.231752 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76rjc" event={"ID":"8cc7a017-bf74-4adb-8333-b504efcae212","Type":"ContainerDied","Data":"eaa5453981b644977fb5bc16093c6e3c139279399ad0f955882da4e798700426"} Oct 07 15:27:46 crc kubenswrapper[4959]: I1007 15:27:46.247195 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76rjc" event={"ID":"8cc7a017-bf74-4adb-8333-b504efcae212","Type":"ContainerStarted","Data":"6ea0ee450bb5de6a01bdca231f72a665eca748d11554280e3a62e5d67eb0c3d8"} Oct 07 15:27:46 crc kubenswrapper[4959]: I1007 15:27:46.294096 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-76rjc" podStartSLOduration=3.579165117 podStartE2EDuration="6.294065772s" podCreationTimestamp="2025-10-07 15:27:40 +0000 UTC" firstStartedPulling="2025-10-07 15:27:43.213445073 +0000 UTC m=+6145.296849398" lastFinishedPulling="2025-10-07 15:27:45.928345728 +0000 UTC m=+6148.011750053" observedRunningTime="2025-10-07 15:27:46.288824018 +0000 UTC m=+6148.372228353" watchObservedRunningTime="2025-10-07 15:27:46.294065772 +0000 UTC m=+6148.377470097" Oct 07 15:27:50 crc kubenswrapper[4959]: I1007 15:27:50.562511 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-76rjc" Oct 07 15:27:50 crc kubenswrapper[4959]: I1007 15:27:50.563086 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-76rjc" Oct 07 15:27:50 crc kubenswrapper[4959]: I1007 15:27:50.627833 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-76rjc" Oct 07 15:27:51 crc kubenswrapper[4959]: I1007 15:27:51.341301 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-76rjc" Oct 07 15:27:51 crc kubenswrapper[4959]: I1007 15:27:51.400066 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-76rjc"] Oct 07 15:27:53 crc kubenswrapper[4959]: I1007 15:27:53.311733 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-76rjc" podUID="8cc7a017-bf74-4adb-8333-b504efcae212" containerName="registry-server" containerID="cri-o://6ea0ee450bb5de6a01bdca231f72a665eca748d11554280e3a62e5d67eb0c3d8" gracePeriod=2 Oct 07 15:27:53 crc kubenswrapper[4959]: I1007 15:27:53.827234 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-76rjc" Oct 07 15:27:53 crc kubenswrapper[4959]: I1007 15:27:53.957611 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cc7a017-bf74-4adb-8333-b504efcae212-catalog-content\") pod \"8cc7a017-bf74-4adb-8333-b504efcae212\" (UID: \"8cc7a017-bf74-4adb-8333-b504efcae212\") " Oct 07 15:27:53 crc kubenswrapper[4959]: I1007 15:27:53.958004 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlvt9\" (UniqueName: \"kubernetes.io/projected/8cc7a017-bf74-4adb-8333-b504efcae212-kube-api-access-tlvt9\") pod \"8cc7a017-bf74-4adb-8333-b504efcae212\" (UID: \"8cc7a017-bf74-4adb-8333-b504efcae212\") " Oct 07 15:27:53 crc kubenswrapper[4959]: I1007 15:27:53.958049 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cc7a017-bf74-4adb-8333-b504efcae212-utilities\") pod \"8cc7a017-bf74-4adb-8333-b504efcae212\" (UID: \"8cc7a017-bf74-4adb-8333-b504efcae212\") " Oct 07 15:27:53 crc kubenswrapper[4959]: I1007 15:27:53.959285 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cc7a017-bf74-4adb-8333-b504efcae212-utilities" (OuterVolumeSpecName: "utilities") pod "8cc7a017-bf74-4adb-8333-b504efcae212" (UID: "8cc7a017-bf74-4adb-8333-b504efcae212"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:27:53 crc kubenswrapper[4959]: I1007 15:27:53.965407 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cc7a017-bf74-4adb-8333-b504efcae212-kube-api-access-tlvt9" (OuterVolumeSpecName: "kube-api-access-tlvt9") pod "8cc7a017-bf74-4adb-8333-b504efcae212" (UID: "8cc7a017-bf74-4adb-8333-b504efcae212"). InnerVolumeSpecName "kube-api-access-tlvt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 15:27:54 crc kubenswrapper[4959]: I1007 15:27:54.013955 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cc7a017-bf74-4adb-8333-b504efcae212-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8cc7a017-bf74-4adb-8333-b504efcae212" (UID: "8cc7a017-bf74-4adb-8333-b504efcae212"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:27:54 crc kubenswrapper[4959]: I1007 15:27:54.060624 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlvt9\" (UniqueName: \"kubernetes.io/projected/8cc7a017-bf74-4adb-8333-b504efcae212-kube-api-access-tlvt9\") on node \"crc\" DevicePath \"\"" Oct 07 15:27:54 crc kubenswrapper[4959]: I1007 15:27:54.060679 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cc7a017-bf74-4adb-8333-b504efcae212-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 15:27:54 crc kubenswrapper[4959]: I1007 15:27:54.060691 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cc7a017-bf74-4adb-8333-b504efcae212-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 15:27:54 crc kubenswrapper[4959]: I1007 15:27:54.325835 4959 generic.go:334] "Generic (PLEG): container finished" podID="8cc7a017-bf74-4adb-8333-b504efcae212" containerID="6ea0ee450bb5de6a01bdca231f72a665eca748d11554280e3a62e5d67eb0c3d8" exitCode=0 Oct 07 15:27:54 crc kubenswrapper[4959]: I1007 15:27:54.325949 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76rjc" event={"ID":"8cc7a017-bf74-4adb-8333-b504efcae212","Type":"ContainerDied","Data":"6ea0ee450bb5de6a01bdca231f72a665eca748d11554280e3a62e5d67eb0c3d8"} Oct 07 15:27:54 crc kubenswrapper[4959]: I1007 15:27:54.326277 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76rjc" event={"ID":"8cc7a017-bf74-4adb-8333-b504efcae212","Type":"ContainerDied","Data":"e3127644ff187993aa02079b4096d463d52dfb7fdc96593e7e69ec11578e7983"} Oct 07 15:27:54 crc kubenswrapper[4959]: I1007 15:27:54.325994 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-76rjc" Oct 07 15:27:54 crc kubenswrapper[4959]: I1007 15:27:54.326318 4959 scope.go:117] "RemoveContainer" containerID="6ea0ee450bb5de6a01bdca231f72a665eca748d11554280e3a62e5d67eb0c3d8" Oct 07 15:27:54 crc kubenswrapper[4959]: I1007 15:27:54.357597 4959 scope.go:117] "RemoveContainer" containerID="eaa5453981b644977fb5bc16093c6e3c139279399ad0f955882da4e798700426" Oct 07 15:27:54 crc kubenswrapper[4959]: I1007 15:27:54.365651 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-76rjc"] Oct 07 15:27:54 crc kubenswrapper[4959]: I1007 15:27:54.375007 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-76rjc"] Oct 07 15:27:54 crc kubenswrapper[4959]: I1007 15:27:54.387325 4959 scope.go:117] "RemoveContainer" containerID="b870cf58fcdfb6ede6e7e75c675720cbe7dbe53f988f7988654b3322bfd990c5" Oct 07 15:27:54 crc kubenswrapper[4959]: I1007 15:27:54.441889 4959 scope.go:117] "RemoveContainer" containerID="6ea0ee450bb5de6a01bdca231f72a665eca748d11554280e3a62e5d67eb0c3d8" Oct 07 15:27:54 crc kubenswrapper[4959]: E1007 15:27:54.442484 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ea0ee450bb5de6a01bdca231f72a665eca748d11554280e3a62e5d67eb0c3d8\": container with ID starting with 6ea0ee450bb5de6a01bdca231f72a665eca748d11554280e3a62e5d67eb0c3d8 not found: ID does not exist" containerID="6ea0ee450bb5de6a01bdca231f72a665eca748d11554280e3a62e5d67eb0c3d8" Oct 07 15:27:54 crc kubenswrapper[4959]: I1007 15:27:54.442558 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ea0ee450bb5de6a01bdca231f72a665eca748d11554280e3a62e5d67eb0c3d8"} err="failed to get container status \"6ea0ee450bb5de6a01bdca231f72a665eca748d11554280e3a62e5d67eb0c3d8\": rpc error: code = NotFound desc = could not find container \"6ea0ee450bb5de6a01bdca231f72a665eca748d11554280e3a62e5d67eb0c3d8\": container with ID starting with 6ea0ee450bb5de6a01bdca231f72a665eca748d11554280e3a62e5d67eb0c3d8 not found: ID does not exist" Oct 07 15:27:54 crc kubenswrapper[4959]: I1007 15:27:54.442599 4959 scope.go:117] "RemoveContainer" containerID="eaa5453981b644977fb5bc16093c6e3c139279399ad0f955882da4e798700426" Oct 07 15:27:54 crc kubenswrapper[4959]: E1007 15:27:54.443076 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eaa5453981b644977fb5bc16093c6e3c139279399ad0f955882da4e798700426\": container with ID starting with eaa5453981b644977fb5bc16093c6e3c139279399ad0f955882da4e798700426 not found: ID does not exist" containerID="eaa5453981b644977fb5bc16093c6e3c139279399ad0f955882da4e798700426" Oct 07 15:27:54 crc kubenswrapper[4959]: I1007 15:27:54.443138 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaa5453981b644977fb5bc16093c6e3c139279399ad0f955882da4e798700426"} err="failed to get container status \"eaa5453981b644977fb5bc16093c6e3c139279399ad0f955882da4e798700426\": rpc error: code = NotFound desc = could not find container \"eaa5453981b644977fb5bc16093c6e3c139279399ad0f955882da4e798700426\": container with ID starting with eaa5453981b644977fb5bc16093c6e3c139279399ad0f955882da4e798700426 not found: ID does not exist" Oct 07 15:27:54 crc kubenswrapper[4959]: I1007 15:27:54.443173 4959 scope.go:117] "RemoveContainer" containerID="b870cf58fcdfb6ede6e7e75c675720cbe7dbe53f988f7988654b3322bfd990c5" Oct 07 15:27:54 crc kubenswrapper[4959]: E1007 15:27:54.443439 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b870cf58fcdfb6ede6e7e75c675720cbe7dbe53f988f7988654b3322bfd990c5\": container with ID starting with b870cf58fcdfb6ede6e7e75c675720cbe7dbe53f988f7988654b3322bfd990c5 not found: ID does not exist" containerID="b870cf58fcdfb6ede6e7e75c675720cbe7dbe53f988f7988654b3322bfd990c5" Oct 07 15:27:54 crc kubenswrapper[4959]: I1007 15:27:54.443471 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b870cf58fcdfb6ede6e7e75c675720cbe7dbe53f988f7988654b3322bfd990c5"} err="failed to get container status \"b870cf58fcdfb6ede6e7e75c675720cbe7dbe53f988f7988654b3322bfd990c5\": rpc error: code = NotFound desc = could not find container \"b870cf58fcdfb6ede6e7e75c675720cbe7dbe53f988f7988654b3322bfd990c5\": container with ID starting with b870cf58fcdfb6ede6e7e75c675720cbe7dbe53f988f7988654b3322bfd990c5 not found: ID does not exist" Oct 07 15:27:54 crc kubenswrapper[4959]: I1007 15:27:54.668054 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cc7a017-bf74-4adb-8333-b504efcae212" path="/var/lib/kubelet/pods/8cc7a017-bf74-4adb-8333-b504efcae212/volumes" Oct 07 15:28:07 crc kubenswrapper[4959]: I1007 15:28:07.371833 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tg9pd"] Oct 07 15:28:07 crc kubenswrapper[4959]: E1007 15:28:07.373677 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cc7a017-bf74-4adb-8333-b504efcae212" containerName="extract-utilities" Oct 07 15:28:07 crc kubenswrapper[4959]: I1007 15:28:07.373697 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cc7a017-bf74-4adb-8333-b504efcae212" containerName="extract-utilities" Oct 07 15:28:07 crc kubenswrapper[4959]: E1007 15:28:07.373755 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cc7a017-bf74-4adb-8333-b504efcae212" containerName="extract-content" Oct 07 15:28:07 crc kubenswrapper[4959]: I1007 15:28:07.373763 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cc7a017-bf74-4adb-8333-b504efcae212" containerName="extract-content" Oct 07 15:28:07 crc kubenswrapper[4959]: E1007 15:28:07.373778 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cc7a017-bf74-4adb-8333-b504efcae212" containerName="registry-server" Oct 07 15:28:07 crc kubenswrapper[4959]: I1007 15:28:07.373784 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cc7a017-bf74-4adb-8333-b504efcae212" containerName="registry-server" Oct 07 15:28:07 crc kubenswrapper[4959]: I1007 15:28:07.374015 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cc7a017-bf74-4adb-8333-b504efcae212" containerName="registry-server" Oct 07 15:28:07 crc kubenswrapper[4959]: I1007 15:28:07.377171 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tg9pd" Oct 07 15:28:07 crc kubenswrapper[4959]: I1007 15:28:07.387044 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tg9pd"] Oct 07 15:28:07 crc kubenswrapper[4959]: I1007 15:28:07.567491 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed-utilities\") pod \"redhat-marketplace-tg9pd\" (UID: \"1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed\") " pod="openshift-marketplace/redhat-marketplace-tg9pd" Oct 07 15:28:07 crc kubenswrapper[4959]: I1007 15:28:07.567659 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vj8kz\" (UniqueName: \"kubernetes.io/projected/1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed-kube-api-access-vj8kz\") pod \"redhat-marketplace-tg9pd\" (UID: \"1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed\") " pod="openshift-marketplace/redhat-marketplace-tg9pd" Oct 07 15:28:07 crc kubenswrapper[4959]: I1007 15:28:07.567884 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed-catalog-content\") pod \"redhat-marketplace-tg9pd\" (UID: \"1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed\") " pod="openshift-marketplace/redhat-marketplace-tg9pd" Oct 07 15:28:07 crc kubenswrapper[4959]: I1007 15:28:07.669635 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed-catalog-content\") pod \"redhat-marketplace-tg9pd\" (UID: \"1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed\") " pod="openshift-marketplace/redhat-marketplace-tg9pd" Oct 07 15:28:07 crc kubenswrapper[4959]: I1007 15:28:07.669789 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed-utilities\") pod \"redhat-marketplace-tg9pd\" (UID: \"1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed\") " pod="openshift-marketplace/redhat-marketplace-tg9pd" Oct 07 15:28:07 crc kubenswrapper[4959]: I1007 15:28:07.669832 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vj8kz\" (UniqueName: \"kubernetes.io/projected/1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed-kube-api-access-vj8kz\") pod \"redhat-marketplace-tg9pd\" (UID: \"1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed\") " pod="openshift-marketplace/redhat-marketplace-tg9pd" Oct 07 15:28:07 crc kubenswrapper[4959]: I1007 15:28:07.670481 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed-catalog-content\") pod \"redhat-marketplace-tg9pd\" (UID: \"1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed\") " pod="openshift-marketplace/redhat-marketplace-tg9pd" Oct 07 15:28:07 crc kubenswrapper[4959]: I1007 15:28:07.670481 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed-utilities\") pod \"redhat-marketplace-tg9pd\" (UID: \"1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed\") " pod="openshift-marketplace/redhat-marketplace-tg9pd" Oct 07 15:28:07 crc kubenswrapper[4959]: I1007 15:28:07.702452 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vj8kz\" (UniqueName: \"kubernetes.io/projected/1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed-kube-api-access-vj8kz\") pod \"redhat-marketplace-tg9pd\" (UID: \"1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed\") " pod="openshift-marketplace/redhat-marketplace-tg9pd" Oct 07 15:28:07 crc kubenswrapper[4959]: I1007 15:28:07.706601 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tg9pd" Oct 07 15:28:08 crc kubenswrapper[4959]: I1007 15:28:08.205672 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tg9pd"] Oct 07 15:28:08 crc kubenswrapper[4959]: I1007 15:28:08.503198 4959 generic.go:334] "Generic (PLEG): container finished" podID="1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed" containerID="7f9a3de5d6dd190716fd06176f6955e84d050ac8b1898706ee185b50ec4f9c50" exitCode=0 Oct 07 15:28:08 crc kubenswrapper[4959]: I1007 15:28:08.503305 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tg9pd" event={"ID":"1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed","Type":"ContainerDied","Data":"7f9a3de5d6dd190716fd06176f6955e84d050ac8b1898706ee185b50ec4f9c50"} Oct 07 15:28:08 crc kubenswrapper[4959]: I1007 15:28:08.503507 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tg9pd" event={"ID":"1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed","Type":"ContainerStarted","Data":"51f6525af640bea6ed87d046791b49fd23228ab654ad3e710ffd033b68b7394e"} Oct 07 15:28:09 crc kubenswrapper[4959]: I1007 15:28:09.520872 4959 generic.go:334] "Generic (PLEG): container finished" podID="1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed" containerID="4e2c1962d5a511c73bd7db3a9468ec194c81e568ae576957f094b813cc79b69e" exitCode=0 Oct 07 15:28:09 crc kubenswrapper[4959]: I1007 15:28:09.521133 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tg9pd" event={"ID":"1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed","Type":"ContainerDied","Data":"4e2c1962d5a511c73bd7db3a9468ec194c81e568ae576957f094b813cc79b69e"} Oct 07 15:28:10 crc kubenswrapper[4959]: I1007 15:28:10.535323 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tg9pd" event={"ID":"1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed","Type":"ContainerStarted","Data":"4bfb9d5df1953da2edc51213a7d986af85b1e4f4f0a4969ab35a6ed32cd43c3e"} Oct 07 15:28:10 crc kubenswrapper[4959]: I1007 15:28:10.573285 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tg9pd" podStartSLOduration=2.110872877 podStartE2EDuration="3.57325715s" podCreationTimestamp="2025-10-07 15:28:07 +0000 UTC" firstStartedPulling="2025-10-07 15:28:08.505052882 +0000 UTC m=+6170.588457207" lastFinishedPulling="2025-10-07 15:28:09.967437155 +0000 UTC m=+6172.050841480" observedRunningTime="2025-10-07 15:28:10.559775486 +0000 UTC m=+6172.643179811" watchObservedRunningTime="2025-10-07 15:28:10.57325715 +0000 UTC m=+6172.656661495" Oct 07 15:28:17 crc kubenswrapper[4959]: I1007 15:28:17.707783 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tg9pd" Oct 07 15:28:17 crc kubenswrapper[4959]: I1007 15:28:17.708382 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tg9pd" Oct 07 15:28:17 crc kubenswrapper[4959]: I1007 15:28:17.760878 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tg9pd" Oct 07 15:28:18 crc kubenswrapper[4959]: I1007 15:28:18.675432 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tg9pd" Oct 07 15:28:18 crc kubenswrapper[4959]: I1007 15:28:18.736655 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tg9pd"] Oct 07 15:28:20 crc kubenswrapper[4959]: I1007 15:28:20.623474 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tg9pd" podUID="1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed" containerName="registry-server" containerID="cri-o://4bfb9d5df1953da2edc51213a7d986af85b1e4f4f0a4969ab35a6ed32cd43c3e" gracePeriod=2 Oct 07 15:28:21 crc kubenswrapper[4959]: I1007 15:28:21.141759 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tg9pd" Oct 07 15:28:21 crc kubenswrapper[4959]: I1007 15:28:21.175632 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed-catalog-content\") pod \"1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed\" (UID: \"1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed\") " Oct 07 15:28:21 crc kubenswrapper[4959]: I1007 15:28:21.175796 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vj8kz\" (UniqueName: \"kubernetes.io/projected/1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed-kube-api-access-vj8kz\") pod \"1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed\" (UID: \"1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed\") " Oct 07 15:28:21 crc kubenswrapper[4959]: I1007 15:28:21.175878 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed-utilities\") pod \"1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed\" (UID: \"1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed\") " Oct 07 15:28:21 crc kubenswrapper[4959]: I1007 15:28:21.180846 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed-utilities" (OuterVolumeSpecName: "utilities") pod "1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed" (UID: "1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:28:21 crc kubenswrapper[4959]: I1007 15:28:21.185441 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed-kube-api-access-vj8kz" (OuterVolumeSpecName: "kube-api-access-vj8kz") pod "1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed" (UID: "1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed"). InnerVolumeSpecName "kube-api-access-vj8kz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 15:28:21 crc kubenswrapper[4959]: I1007 15:28:21.191712 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed" (UID: "1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:28:21 crc kubenswrapper[4959]: I1007 15:28:21.278862 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 15:28:21 crc kubenswrapper[4959]: I1007 15:28:21.278898 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 15:28:21 crc kubenswrapper[4959]: I1007 15:28:21.278911 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vj8kz\" (UniqueName: \"kubernetes.io/projected/1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed-kube-api-access-vj8kz\") on node \"crc\" DevicePath \"\"" Oct 07 15:28:21 crc kubenswrapper[4959]: I1007 15:28:21.637288 4959 generic.go:334] "Generic (PLEG): container finished" podID="1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed" containerID="4bfb9d5df1953da2edc51213a7d986af85b1e4f4f0a4969ab35a6ed32cd43c3e" exitCode=0 Oct 07 15:28:21 crc kubenswrapper[4959]: I1007 15:28:21.637344 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tg9pd" event={"ID":"1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed","Type":"ContainerDied","Data":"4bfb9d5df1953da2edc51213a7d986af85b1e4f4f0a4969ab35a6ed32cd43c3e"} Oct 07 15:28:21 crc kubenswrapper[4959]: I1007 15:28:21.637404 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tg9pd" event={"ID":"1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed","Type":"ContainerDied","Data":"51f6525af640bea6ed87d046791b49fd23228ab654ad3e710ffd033b68b7394e"} Oct 07 15:28:21 crc kubenswrapper[4959]: I1007 15:28:21.637455 4959 scope.go:117] "RemoveContainer" containerID="4bfb9d5df1953da2edc51213a7d986af85b1e4f4f0a4969ab35a6ed32cd43c3e" Oct 07 15:28:21 crc kubenswrapper[4959]: I1007 15:28:21.639054 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tg9pd" Oct 07 15:28:21 crc kubenswrapper[4959]: I1007 15:28:21.672571 4959 scope.go:117] "RemoveContainer" containerID="4e2c1962d5a511c73bd7db3a9468ec194c81e568ae576957f094b813cc79b69e" Oct 07 15:28:21 crc kubenswrapper[4959]: I1007 15:28:21.678482 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tg9pd"] Oct 07 15:28:21 crc kubenswrapper[4959]: I1007 15:28:21.697009 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tg9pd"] Oct 07 15:28:21 crc kubenswrapper[4959]: I1007 15:28:21.699895 4959 scope.go:117] "RemoveContainer" containerID="7f9a3de5d6dd190716fd06176f6955e84d050ac8b1898706ee185b50ec4f9c50" Oct 07 15:28:21 crc kubenswrapper[4959]: I1007 15:28:21.748856 4959 scope.go:117] "RemoveContainer" containerID="4bfb9d5df1953da2edc51213a7d986af85b1e4f4f0a4969ab35a6ed32cd43c3e" Oct 07 15:28:21 crc kubenswrapper[4959]: E1007 15:28:21.749609 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bfb9d5df1953da2edc51213a7d986af85b1e4f4f0a4969ab35a6ed32cd43c3e\": container with ID starting with 4bfb9d5df1953da2edc51213a7d986af85b1e4f4f0a4969ab35a6ed32cd43c3e not found: ID does not exist" containerID="4bfb9d5df1953da2edc51213a7d986af85b1e4f4f0a4969ab35a6ed32cd43c3e" Oct 07 15:28:21 crc kubenswrapper[4959]: I1007 15:28:21.749681 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bfb9d5df1953da2edc51213a7d986af85b1e4f4f0a4969ab35a6ed32cd43c3e"} err="failed to get container status \"4bfb9d5df1953da2edc51213a7d986af85b1e4f4f0a4969ab35a6ed32cd43c3e\": rpc error: code = NotFound desc = could not find container \"4bfb9d5df1953da2edc51213a7d986af85b1e4f4f0a4969ab35a6ed32cd43c3e\": container with ID starting with 4bfb9d5df1953da2edc51213a7d986af85b1e4f4f0a4969ab35a6ed32cd43c3e not found: ID does not exist" Oct 07 15:28:21 crc kubenswrapper[4959]: I1007 15:28:21.749731 4959 scope.go:117] "RemoveContainer" containerID="4e2c1962d5a511c73bd7db3a9468ec194c81e568ae576957f094b813cc79b69e" Oct 07 15:28:21 crc kubenswrapper[4959]: E1007 15:28:21.750266 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e2c1962d5a511c73bd7db3a9468ec194c81e568ae576957f094b813cc79b69e\": container with ID starting with 4e2c1962d5a511c73bd7db3a9468ec194c81e568ae576957f094b813cc79b69e not found: ID does not exist" containerID="4e2c1962d5a511c73bd7db3a9468ec194c81e568ae576957f094b813cc79b69e" Oct 07 15:28:21 crc kubenswrapper[4959]: I1007 15:28:21.750318 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e2c1962d5a511c73bd7db3a9468ec194c81e568ae576957f094b813cc79b69e"} err="failed to get container status \"4e2c1962d5a511c73bd7db3a9468ec194c81e568ae576957f094b813cc79b69e\": rpc error: code = NotFound desc = could not find container \"4e2c1962d5a511c73bd7db3a9468ec194c81e568ae576957f094b813cc79b69e\": container with ID starting with 4e2c1962d5a511c73bd7db3a9468ec194c81e568ae576957f094b813cc79b69e not found: ID does not exist" Oct 07 15:28:21 crc kubenswrapper[4959]: I1007 15:28:21.750352 4959 scope.go:117] "RemoveContainer" containerID="7f9a3de5d6dd190716fd06176f6955e84d050ac8b1898706ee185b50ec4f9c50" Oct 07 15:28:21 crc kubenswrapper[4959]: E1007 15:28:21.751064 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f9a3de5d6dd190716fd06176f6955e84d050ac8b1898706ee185b50ec4f9c50\": container with ID starting with 7f9a3de5d6dd190716fd06176f6955e84d050ac8b1898706ee185b50ec4f9c50 not found: ID does not exist" containerID="7f9a3de5d6dd190716fd06176f6955e84d050ac8b1898706ee185b50ec4f9c50" Oct 07 15:28:21 crc kubenswrapper[4959]: I1007 15:28:21.751575 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f9a3de5d6dd190716fd06176f6955e84d050ac8b1898706ee185b50ec4f9c50"} err="failed to get container status \"7f9a3de5d6dd190716fd06176f6955e84d050ac8b1898706ee185b50ec4f9c50\": rpc error: code = NotFound desc = could not find container \"7f9a3de5d6dd190716fd06176f6955e84d050ac8b1898706ee185b50ec4f9c50\": container with ID starting with 7f9a3de5d6dd190716fd06176f6955e84d050ac8b1898706ee185b50ec4f9c50 not found: ID does not exist" Oct 07 15:28:22 crc kubenswrapper[4959]: I1007 15:28:22.666031 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed" path="/var/lib/kubelet/pods/1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed/volumes" Oct 07 15:28:50 crc kubenswrapper[4959]: I1007 15:28:50.630154 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 15:28:50 crc kubenswrapper[4959]: I1007 15:28:50.630820 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 15:29:03 crc kubenswrapper[4959]: I1007 15:29:03.011982 4959 generic.go:334] "Generic (PLEG): container finished" podID="a646a7f2-705e-4278-9950-1a20b2db0149" containerID="e8d188e7a4882290cbeb5076f525e129320a29fe0604137f84342cb01782be2b" exitCode=0 Oct 07 15:29:03 crc kubenswrapper[4959]: I1007 15:29:03.012078 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" event={"ID":"a646a7f2-705e-4278-9950-1a20b2db0149","Type":"ContainerDied","Data":"e8d188e7a4882290cbeb5076f525e129320a29fe0604137f84342cb01782be2b"} Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.422777 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.491758 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tobiko-tests-tobiko-s01-sanity"] Oct 07 15:29:04 crc kubenswrapper[4959]: E1007 15:29:04.492211 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed" containerName="extract-utilities" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.492229 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed" containerName="extract-utilities" Oct 07 15:29:04 crc kubenswrapper[4959]: E1007 15:29:04.492276 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a646a7f2-705e-4278-9950-1a20b2db0149" containerName="tobiko-tests-tobiko" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.492285 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a646a7f2-705e-4278-9950-1a20b2db0149" containerName="tobiko-tests-tobiko" Oct 07 15:29:04 crc kubenswrapper[4959]: E1007 15:29:04.492320 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed" containerName="registry-server" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.492326 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed" containerName="registry-server" Oct 07 15:29:04 crc kubenswrapper[4959]: E1007 15:29:04.492339 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed" containerName="extract-content" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.492345 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed" containerName="extract-content" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.492543 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a646a7f2-705e-4278-9950-1a20b2db0149" containerName="tobiko-tests-tobiko" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.492564 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f6ec79e-0ebe-4549-9cf4-0a695b6ba5ed" containerName="registry-server" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.493220 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.507986 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tobiko-tests-tobiko-s01-sanity"] Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.531601 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a646a7f2-705e-4278-9950-1a20b2db0149-ceph\") pod \"a646a7f2-705e-4278-9950-1a20b2db0149\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.531737 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/a646a7f2-705e-4278-9950-1a20b2db0149-test-operator-clouds-config\") pod \"a646a7f2-705e-4278-9950-1a20b2db0149\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.531935 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a646a7f2-705e-4278-9950-1a20b2db0149-ca-certs\") pod \"a646a7f2-705e-4278-9950-1a20b2db0149\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.531993 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/a646a7f2-705e-4278-9950-1a20b2db0149-kubeconfig\") pod \"a646a7f2-705e-4278-9950-1a20b2db0149\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.532093 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a646a7f2-705e-4278-9950-1a20b2db0149-test-operator-ephemeral-workdir\") pod \"a646a7f2-705e-4278-9950-1a20b2db0149\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.532190 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/a646a7f2-705e-4278-9950-1a20b2db0149-tobiko-private-key\") pod \"a646a7f2-705e-4278-9950-1a20b2db0149\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.532231 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/a646a7f2-705e-4278-9950-1a20b2db0149-tobiko-config\") pod \"a646a7f2-705e-4278-9950-1a20b2db0149\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.532261 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/a646a7f2-705e-4278-9950-1a20b2db0149-tobiko-public-key\") pod \"a646a7f2-705e-4278-9950-1a20b2db0149\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.532385 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a646a7f2-705e-4278-9950-1a20b2db0149-openstack-config-secret\") pod \"a646a7f2-705e-4278-9950-1a20b2db0149\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.532450 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"a646a7f2-705e-4278-9950-1a20b2db0149\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.532510 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a646a7f2-705e-4278-9950-1a20b2db0149-test-operator-ephemeral-temporary\") pod \"a646a7f2-705e-4278-9950-1a20b2db0149\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.532553 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8kt5\" (UniqueName: \"kubernetes.io/projected/a646a7f2-705e-4278-9950-1a20b2db0149-kube-api-access-f8kt5\") pod \"a646a7f2-705e-4278-9950-1a20b2db0149\" (UID: \"a646a7f2-705e-4278-9950-1a20b2db0149\") " Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.537158 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a646a7f2-705e-4278-9950-1a20b2db0149-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "a646a7f2-705e-4278-9950-1a20b2db0149" (UID: "a646a7f2-705e-4278-9950-1a20b2db0149"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.538494 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a646a7f2-705e-4278-9950-1a20b2db0149-ceph" (OuterVolumeSpecName: "ceph") pod "a646a7f2-705e-4278-9950-1a20b2db0149" (UID: "a646a7f2-705e-4278-9950-1a20b2db0149"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.556432 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a646a7f2-705e-4278-9950-1a20b2db0149-kube-api-access-f8kt5" (OuterVolumeSpecName: "kube-api-access-f8kt5") pod "a646a7f2-705e-4278-9950-1a20b2db0149" (UID: "a646a7f2-705e-4278-9950-1a20b2db0149"). InnerVolumeSpecName "kube-api-access-f8kt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.560440 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "test-operator-logs") pod "a646a7f2-705e-4278-9950-1a20b2db0149" (UID: "a646a7f2-705e-4278-9950-1a20b2db0149"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.563557 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a646a7f2-705e-4278-9950-1a20b2db0149-tobiko-public-key" (OuterVolumeSpecName: "tobiko-public-key") pod "a646a7f2-705e-4278-9950-1a20b2db0149" (UID: "a646a7f2-705e-4278-9950-1a20b2db0149"). InnerVolumeSpecName "tobiko-public-key". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.565941 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a646a7f2-705e-4278-9950-1a20b2db0149-tobiko-private-key" (OuterVolumeSpecName: "tobiko-private-key") pod "a646a7f2-705e-4278-9950-1a20b2db0149" (UID: "a646a7f2-705e-4278-9950-1a20b2db0149"). InnerVolumeSpecName "tobiko-private-key". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.570509 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a646a7f2-705e-4278-9950-1a20b2db0149-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "a646a7f2-705e-4278-9950-1a20b2db0149" (UID: "a646a7f2-705e-4278-9950-1a20b2db0149"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.576593 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a646a7f2-705e-4278-9950-1a20b2db0149-kubeconfig" (OuterVolumeSpecName: "kubeconfig") pod "a646a7f2-705e-4278-9950-1a20b2db0149" (UID: "a646a7f2-705e-4278-9950-1a20b2db0149"). InnerVolumeSpecName "kubeconfig". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.583824 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a646a7f2-705e-4278-9950-1a20b2db0149-tobiko-config" (OuterVolumeSpecName: "tobiko-config") pod "a646a7f2-705e-4278-9950-1a20b2db0149" (UID: "a646a7f2-705e-4278-9950-1a20b2db0149"). InnerVolumeSpecName "tobiko-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.598743 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a646a7f2-705e-4278-9950-1a20b2db0149-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "a646a7f2-705e-4278-9950-1a20b2db0149" (UID: "a646a7f2-705e-4278-9950-1a20b2db0149"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.600388 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a646a7f2-705e-4278-9950-1a20b2db0149-test-operator-clouds-config" (OuterVolumeSpecName: "test-operator-clouds-config") pod "a646a7f2-705e-4278-9950-1a20b2db0149" (UID: "a646a7f2-705e-4278-9950-1a20b2db0149"). InnerVolumeSpecName "test-operator-clouds-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.635313 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/11e9b8da-1cca-4958-8838-4e7fc1364a78-ca-certs\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.635375 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/11e9b8da-1cca-4958-8838-4e7fc1364a78-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.635448 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/11e9b8da-1cca-4958-8838-4e7fc1364a78-tobiko-public-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.635478 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/11e9b8da-1cca-4958-8838-4e7fc1364a78-tobiko-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.635501 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/11e9b8da-1cca-4958-8838-4e7fc1364a78-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.635536 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/11e9b8da-1cca-4958-8838-4e7fc1364a78-tobiko-private-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.635551 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtft9\" (UniqueName: \"kubernetes.io/projected/11e9b8da-1cca-4958-8838-4e7fc1364a78-kube-api-access-rtft9\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.635606 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/11e9b8da-1cca-4958-8838-4e7fc1364a78-kubeconfig\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.635638 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/11e9b8da-1cca-4958-8838-4e7fc1364a78-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.635660 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/11e9b8da-1cca-4958-8838-4e7fc1364a78-ceph\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.635695 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/11e9b8da-1cca-4958-8838-4e7fc1364a78-openstack-config-secret\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.635722 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.635781 4959 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a646a7f2-705e-4278-9950-1a20b2db0149-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.635791 4959 reconciler_common.go:293] "Volume detached for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/a646a7f2-705e-4278-9950-1a20b2db0149-kubeconfig\") on node \"crc\" DevicePath \"\"" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.635801 4959 reconciler_common.go:293] "Volume detached for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/a646a7f2-705e-4278-9950-1a20b2db0149-tobiko-private-key\") on node \"crc\" DevicePath \"\"" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.635811 4959 reconciler_common.go:293] "Volume detached for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/a646a7f2-705e-4278-9950-1a20b2db0149-tobiko-config\") on node \"crc\" DevicePath \"\"" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.635819 4959 reconciler_common.go:293] "Volume detached for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/a646a7f2-705e-4278-9950-1a20b2db0149-tobiko-public-key\") on node \"crc\" DevicePath \"\"" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.635827 4959 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a646a7f2-705e-4278-9950-1a20b2db0149-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.635836 4959 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a646a7f2-705e-4278-9950-1a20b2db0149-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.635844 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8kt5\" (UniqueName: \"kubernetes.io/projected/a646a7f2-705e-4278-9950-1a20b2db0149-kube-api-access-f8kt5\") on node \"crc\" DevicePath \"\"" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.635856 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a646a7f2-705e-4278-9950-1a20b2db0149-ceph\") on node \"crc\" DevicePath \"\"" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.635864 4959 reconciler_common.go:293] "Volume detached for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/a646a7f2-705e-4278-9950-1a20b2db0149-test-operator-clouds-config\") on node \"crc\" DevicePath \"\"" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.697888 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.737673 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/11e9b8da-1cca-4958-8838-4e7fc1364a78-tobiko-public-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.737759 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/11e9b8da-1cca-4958-8838-4e7fc1364a78-tobiko-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.737795 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/11e9b8da-1cca-4958-8838-4e7fc1364a78-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.737839 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/11e9b8da-1cca-4958-8838-4e7fc1364a78-tobiko-private-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.737859 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtft9\" (UniqueName: \"kubernetes.io/projected/11e9b8da-1cca-4958-8838-4e7fc1364a78-kube-api-access-rtft9\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.737975 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/11e9b8da-1cca-4958-8838-4e7fc1364a78-kubeconfig\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.738016 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/11e9b8da-1cca-4958-8838-4e7fc1364a78-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.738043 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/11e9b8da-1cca-4958-8838-4e7fc1364a78-ceph\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.738114 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/11e9b8da-1cca-4958-8838-4e7fc1364a78-openstack-config-secret\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.738160 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/11e9b8da-1cca-4958-8838-4e7fc1364a78-ca-certs\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.738215 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/11e9b8da-1cca-4958-8838-4e7fc1364a78-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.738988 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/11e9b8da-1cca-4958-8838-4e7fc1364a78-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.738995 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/11e9b8da-1cca-4958-8838-4e7fc1364a78-tobiko-public-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.741363 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/11e9b8da-1cca-4958-8838-4e7fc1364a78-tobiko-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.741483 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/11e9b8da-1cca-4958-8838-4e7fc1364a78-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.742371 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/11e9b8da-1cca-4958-8838-4e7fc1364a78-tobiko-private-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.742576 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/11e9b8da-1cca-4958-8838-4e7fc1364a78-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.743665 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/11e9b8da-1cca-4958-8838-4e7fc1364a78-kubeconfig\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.744706 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/11e9b8da-1cca-4958-8838-4e7fc1364a78-openstack-config-secret\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.746993 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/11e9b8da-1cca-4958-8838-4e7fc1364a78-ceph\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.747176 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/11e9b8da-1cca-4958-8838-4e7fc1364a78-ca-certs\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.757009 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtft9\" (UniqueName: \"kubernetes.io/projected/11e9b8da-1cca-4958-8838-4e7fc1364a78-kube-api-access-rtft9\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:04 crc kubenswrapper[4959]: I1007 15:29:04.822565 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:29:05 crc kubenswrapper[4959]: I1007 15:29:05.039850 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" event={"ID":"a646a7f2-705e-4278-9950-1a20b2db0149","Type":"ContainerDied","Data":"5c76f28c5507de82b2666bf23c4cffaf6c42767cd17fa9e5b87bff4f143dd524"} Oct 07 15:29:05 crc kubenswrapper[4959]: I1007 15:29:05.040255 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c76f28c5507de82b2666bf23c4cffaf6c42767cd17fa9e5b87bff4f143dd524" Oct 07 15:29:05 crc kubenswrapper[4959]: I1007 15:29:05.039939 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 07 15:29:05 crc kubenswrapper[4959]: I1007 15:29:05.359209 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tobiko-tests-tobiko-s01-sanity"] Oct 07 15:29:06 crc kubenswrapper[4959]: I1007 15:29:06.047913 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a646a7f2-705e-4278-9950-1a20b2db0149-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "a646a7f2-705e-4278-9950-1a20b2db0149" (UID: "a646a7f2-705e-4278-9950-1a20b2db0149"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:29:06 crc kubenswrapper[4959]: I1007 15:29:06.053593 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s01-sanity" event={"ID":"11e9b8da-1cca-4958-8838-4e7fc1364a78","Type":"ContainerStarted","Data":"77e556549e4d10304225ba03fd689a4455445c0ba2aaf8248ee4afc1fcfdb148"} Oct 07 15:29:06 crc kubenswrapper[4959]: I1007 15:29:06.069496 4959 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a646a7f2-705e-4278-9950-1a20b2db0149-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 07 15:29:07 crc kubenswrapper[4959]: I1007 15:29:07.069967 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s01-sanity" event={"ID":"11e9b8da-1cca-4958-8838-4e7fc1364a78","Type":"ContainerStarted","Data":"1af461d042e48c01ccab257b6f4d7e820fd38ed86601f44788838058d418a310"} Oct 07 15:29:07 crc kubenswrapper[4959]: I1007 15:29:07.087279 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tobiko-tests-tobiko-s01-sanity" podStartSLOduration=3.087256109 podStartE2EDuration="3.087256109s" podCreationTimestamp="2025-10-07 15:29:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-07 15:29:07.085659968 +0000 UTC m=+6229.169064303" watchObservedRunningTime="2025-10-07 15:29:07.087256109 +0000 UTC m=+6229.170660434" Oct 07 15:29:20 crc kubenswrapper[4959]: I1007 15:29:20.630194 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 15:29:20 crc kubenswrapper[4959]: I1007 15:29:20.631201 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 15:29:50 crc kubenswrapper[4959]: I1007 15:29:50.629972 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 15:29:50 crc kubenswrapper[4959]: I1007 15:29:50.630535 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 15:29:50 crc kubenswrapper[4959]: I1007 15:29:50.630585 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 15:29:50 crc kubenswrapper[4959]: I1007 15:29:50.631360 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3b696f2313816349c23c9e18632aa1ab844c286a6de0e092da7dc16410d9a66e"} pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 15:29:50 crc kubenswrapper[4959]: I1007 15:29:50.631418 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" containerID="cri-o://3b696f2313816349c23c9e18632aa1ab844c286a6de0e092da7dc16410d9a66e" gracePeriod=600 Oct 07 15:29:51 crc kubenswrapper[4959]: I1007 15:29:51.487357 4959 generic.go:334] "Generic (PLEG): container finished" podID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerID="3b696f2313816349c23c9e18632aa1ab844c286a6de0e092da7dc16410d9a66e" exitCode=0 Oct 07 15:29:51 crc kubenswrapper[4959]: I1007 15:29:51.488015 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerDied","Data":"3b696f2313816349c23c9e18632aa1ab844c286a6de0e092da7dc16410d9a66e"} Oct 07 15:29:51 crc kubenswrapper[4959]: I1007 15:29:51.488075 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerStarted","Data":"02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a"} Oct 07 15:29:51 crc kubenswrapper[4959]: I1007 15:29:51.488135 4959 scope.go:117] "RemoveContainer" containerID="20bb9db4c5530cd863be29eafb8adf2f39ef970fbb6588187d2d15050ef33168" Oct 07 15:30:00 crc kubenswrapper[4959]: I1007 15:30:00.175052 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330850-b86ms"] Oct 07 15:30:00 crc kubenswrapper[4959]: I1007 15:30:00.176974 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330850-b86ms" Oct 07 15:30:00 crc kubenswrapper[4959]: I1007 15:30:00.179759 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 07 15:30:00 crc kubenswrapper[4959]: I1007 15:30:00.179999 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 07 15:30:00 crc kubenswrapper[4959]: I1007 15:30:00.195572 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330850-b86ms"] Oct 07 15:30:00 crc kubenswrapper[4959]: I1007 15:30:00.346995 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df241fc3-4163-4d9a-9b82-af3bc490dd85-config-volume\") pod \"collect-profiles-29330850-b86ms\" (UID: \"df241fc3-4163-4d9a-9b82-af3bc490dd85\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330850-b86ms" Oct 07 15:30:00 crc kubenswrapper[4959]: I1007 15:30:00.347084 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m7sg\" (UniqueName: \"kubernetes.io/projected/df241fc3-4163-4d9a-9b82-af3bc490dd85-kube-api-access-7m7sg\") pod \"collect-profiles-29330850-b86ms\" (UID: \"df241fc3-4163-4d9a-9b82-af3bc490dd85\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330850-b86ms" Oct 07 15:30:00 crc kubenswrapper[4959]: I1007 15:30:00.347282 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df241fc3-4163-4d9a-9b82-af3bc490dd85-secret-volume\") pod \"collect-profiles-29330850-b86ms\" (UID: \"df241fc3-4163-4d9a-9b82-af3bc490dd85\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330850-b86ms" Oct 07 15:30:00 crc kubenswrapper[4959]: I1007 15:30:00.449410 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df241fc3-4163-4d9a-9b82-af3bc490dd85-secret-volume\") pod \"collect-profiles-29330850-b86ms\" (UID: \"df241fc3-4163-4d9a-9b82-af3bc490dd85\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330850-b86ms" Oct 07 15:30:00 crc kubenswrapper[4959]: I1007 15:30:00.449653 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df241fc3-4163-4d9a-9b82-af3bc490dd85-config-volume\") pod \"collect-profiles-29330850-b86ms\" (UID: \"df241fc3-4163-4d9a-9b82-af3bc490dd85\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330850-b86ms" Oct 07 15:30:00 crc kubenswrapper[4959]: I1007 15:30:00.449676 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m7sg\" (UniqueName: \"kubernetes.io/projected/df241fc3-4163-4d9a-9b82-af3bc490dd85-kube-api-access-7m7sg\") pod \"collect-profiles-29330850-b86ms\" (UID: \"df241fc3-4163-4d9a-9b82-af3bc490dd85\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330850-b86ms" Oct 07 15:30:00 crc kubenswrapper[4959]: I1007 15:30:00.451744 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df241fc3-4163-4d9a-9b82-af3bc490dd85-config-volume\") pod \"collect-profiles-29330850-b86ms\" (UID: \"df241fc3-4163-4d9a-9b82-af3bc490dd85\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330850-b86ms" Oct 07 15:30:00 crc kubenswrapper[4959]: I1007 15:30:00.457862 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df241fc3-4163-4d9a-9b82-af3bc490dd85-secret-volume\") pod \"collect-profiles-29330850-b86ms\" (UID: \"df241fc3-4163-4d9a-9b82-af3bc490dd85\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330850-b86ms" Oct 07 15:30:00 crc kubenswrapper[4959]: I1007 15:30:00.471469 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m7sg\" (UniqueName: \"kubernetes.io/projected/df241fc3-4163-4d9a-9b82-af3bc490dd85-kube-api-access-7m7sg\") pod \"collect-profiles-29330850-b86ms\" (UID: \"df241fc3-4163-4d9a-9b82-af3bc490dd85\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330850-b86ms" Oct 07 15:30:00 crc kubenswrapper[4959]: I1007 15:30:00.499790 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330850-b86ms" Oct 07 15:30:00 crc kubenswrapper[4959]: I1007 15:30:00.965883 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330850-b86ms"] Oct 07 15:30:01 crc kubenswrapper[4959]: I1007 15:30:01.587798 4959 generic.go:334] "Generic (PLEG): container finished" podID="df241fc3-4163-4d9a-9b82-af3bc490dd85" containerID="2799a1073f1fc0b8c8f241bfb529083ead126a4e0781d026a6ae75d6f35e675d" exitCode=0 Oct 07 15:30:01 crc kubenswrapper[4959]: I1007 15:30:01.587867 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330850-b86ms" event={"ID":"df241fc3-4163-4d9a-9b82-af3bc490dd85","Type":"ContainerDied","Data":"2799a1073f1fc0b8c8f241bfb529083ead126a4e0781d026a6ae75d6f35e675d"} Oct 07 15:30:01 crc kubenswrapper[4959]: I1007 15:30:01.588219 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330850-b86ms" event={"ID":"df241fc3-4163-4d9a-9b82-af3bc490dd85","Type":"ContainerStarted","Data":"6e6845383e3680e9d8abf5c8e758f3976cf8ea456a3f92cb5ab3703b6d731593"} Oct 07 15:30:03 crc kubenswrapper[4959]: I1007 15:30:03.029266 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330850-b86ms" Oct 07 15:30:03 crc kubenswrapper[4959]: I1007 15:30:03.209640 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df241fc3-4163-4d9a-9b82-af3bc490dd85-config-volume\") pod \"df241fc3-4163-4d9a-9b82-af3bc490dd85\" (UID: \"df241fc3-4163-4d9a-9b82-af3bc490dd85\") " Oct 07 15:30:03 crc kubenswrapper[4959]: I1007 15:30:03.209719 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df241fc3-4163-4d9a-9b82-af3bc490dd85-secret-volume\") pod \"df241fc3-4163-4d9a-9b82-af3bc490dd85\" (UID: \"df241fc3-4163-4d9a-9b82-af3bc490dd85\") " Oct 07 15:30:03 crc kubenswrapper[4959]: I1007 15:30:03.209936 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7m7sg\" (UniqueName: \"kubernetes.io/projected/df241fc3-4163-4d9a-9b82-af3bc490dd85-kube-api-access-7m7sg\") pod \"df241fc3-4163-4d9a-9b82-af3bc490dd85\" (UID: \"df241fc3-4163-4d9a-9b82-af3bc490dd85\") " Oct 07 15:30:03 crc kubenswrapper[4959]: I1007 15:30:03.210612 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df241fc3-4163-4d9a-9b82-af3bc490dd85-config-volume" (OuterVolumeSpecName: "config-volume") pod "df241fc3-4163-4d9a-9b82-af3bc490dd85" (UID: "df241fc3-4163-4d9a-9b82-af3bc490dd85"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 15:30:03 crc kubenswrapper[4959]: I1007 15:30:03.219985 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df241fc3-4163-4d9a-9b82-af3bc490dd85-kube-api-access-7m7sg" (OuterVolumeSpecName: "kube-api-access-7m7sg") pod "df241fc3-4163-4d9a-9b82-af3bc490dd85" (UID: "df241fc3-4163-4d9a-9b82-af3bc490dd85"). InnerVolumeSpecName "kube-api-access-7m7sg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 15:30:03 crc kubenswrapper[4959]: I1007 15:30:03.220548 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df241fc3-4163-4d9a-9b82-af3bc490dd85-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "df241fc3-4163-4d9a-9b82-af3bc490dd85" (UID: "df241fc3-4163-4d9a-9b82-af3bc490dd85"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:30:03 crc kubenswrapper[4959]: I1007 15:30:03.312471 4959 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df241fc3-4163-4d9a-9b82-af3bc490dd85-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 15:30:03 crc kubenswrapper[4959]: I1007 15:30:03.312788 4959 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df241fc3-4163-4d9a-9b82-af3bc490dd85-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 07 15:30:03 crc kubenswrapper[4959]: I1007 15:30:03.312813 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7m7sg\" (UniqueName: \"kubernetes.io/projected/df241fc3-4163-4d9a-9b82-af3bc490dd85-kube-api-access-7m7sg\") on node \"crc\" DevicePath \"\"" Oct 07 15:30:03 crc kubenswrapper[4959]: I1007 15:30:03.609396 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330850-b86ms" event={"ID":"df241fc3-4163-4d9a-9b82-af3bc490dd85","Type":"ContainerDied","Data":"6e6845383e3680e9d8abf5c8e758f3976cf8ea456a3f92cb5ab3703b6d731593"} Oct 07 15:30:03 crc kubenswrapper[4959]: I1007 15:30:03.609452 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e6845383e3680e9d8abf5c8e758f3976cf8ea456a3f92cb5ab3703b6d731593" Oct 07 15:30:03 crc kubenswrapper[4959]: I1007 15:30:03.609544 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330850-b86ms" Oct 07 15:30:04 crc kubenswrapper[4959]: I1007 15:30:04.111831 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330805-d5wdj"] Oct 07 15:30:04 crc kubenswrapper[4959]: I1007 15:30:04.120881 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330805-d5wdj"] Oct 07 15:30:04 crc kubenswrapper[4959]: I1007 15:30:04.682286 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="097894fe-84e7-4e3b-a7cc-2f8c84026a41" path="/var/lib/kubelet/pods/097894fe-84e7-4e3b-a7cc-2f8c84026a41/volumes" Oct 07 15:30:36 crc kubenswrapper[4959]: I1007 15:30:36.554426 4959 scope.go:117] "RemoveContainer" containerID="21836ce70b60d0a99fa88c251b4a664d70936373bfd9df420d01b00e91d558e1" Oct 07 15:30:55 crc kubenswrapper[4959]: I1007 15:30:55.118420 4959 generic.go:334] "Generic (PLEG): container finished" podID="11e9b8da-1cca-4958-8838-4e7fc1364a78" containerID="1af461d042e48c01ccab257b6f4d7e820fd38ed86601f44788838058d418a310" exitCode=0 Oct 07 15:30:55 crc kubenswrapper[4959]: I1007 15:30:55.118517 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s01-sanity" event={"ID":"11e9b8da-1cca-4958-8838-4e7fc1364a78","Type":"ContainerDied","Data":"1af461d042e48c01ccab257b6f4d7e820fd38ed86601f44788838058d418a310"} Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.630133 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.732457 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/11e9b8da-1cca-4958-8838-4e7fc1364a78-test-operator-ephemeral-temporary\") pod \"11e9b8da-1cca-4958-8838-4e7fc1364a78\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.732514 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/11e9b8da-1cca-4958-8838-4e7fc1364a78-ceph\") pod \"11e9b8da-1cca-4958-8838-4e7fc1364a78\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.732551 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/11e9b8da-1cca-4958-8838-4e7fc1364a78-tobiko-public-key\") pod \"11e9b8da-1cca-4958-8838-4e7fc1364a78\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.732584 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/11e9b8da-1cca-4958-8838-4e7fc1364a78-tobiko-config\") pod \"11e9b8da-1cca-4958-8838-4e7fc1364a78\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.732658 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/11e9b8da-1cca-4958-8838-4e7fc1364a78-openstack-config-secret\") pod \"11e9b8da-1cca-4958-8838-4e7fc1364a78\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.732711 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/11e9b8da-1cca-4958-8838-4e7fc1364a78-tobiko-private-key\") pod \"11e9b8da-1cca-4958-8838-4e7fc1364a78\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.732745 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/11e9b8da-1cca-4958-8838-4e7fc1364a78-kubeconfig\") pod \"11e9b8da-1cca-4958-8838-4e7fc1364a78\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.732761 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtft9\" (UniqueName: \"kubernetes.io/projected/11e9b8da-1cca-4958-8838-4e7fc1364a78-kube-api-access-rtft9\") pod \"11e9b8da-1cca-4958-8838-4e7fc1364a78\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.732815 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/11e9b8da-1cca-4958-8838-4e7fc1364a78-test-operator-clouds-config\") pod \"11e9b8da-1cca-4958-8838-4e7fc1364a78\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.732876 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/11e9b8da-1cca-4958-8838-4e7fc1364a78-ca-certs\") pod \"11e9b8da-1cca-4958-8838-4e7fc1364a78\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.732920 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/11e9b8da-1cca-4958-8838-4e7fc1364a78-test-operator-ephemeral-workdir\") pod \"11e9b8da-1cca-4958-8838-4e7fc1364a78\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.732952 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"11e9b8da-1cca-4958-8838-4e7fc1364a78\" (UID: \"11e9b8da-1cca-4958-8838-4e7fc1364a78\") " Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.735260 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11e9b8da-1cca-4958-8838-4e7fc1364a78-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "11e9b8da-1cca-4958-8838-4e7fc1364a78" (UID: "11e9b8da-1cca-4958-8838-4e7fc1364a78"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.740209 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11e9b8da-1cca-4958-8838-4e7fc1364a78-kube-api-access-rtft9" (OuterVolumeSpecName: "kube-api-access-rtft9") pod "11e9b8da-1cca-4958-8838-4e7fc1364a78" (UID: "11e9b8da-1cca-4958-8838-4e7fc1364a78"). InnerVolumeSpecName "kube-api-access-rtft9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.740678 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11e9b8da-1cca-4958-8838-4e7fc1364a78-ceph" (OuterVolumeSpecName: "ceph") pod "11e9b8da-1cca-4958-8838-4e7fc1364a78" (UID: "11e9b8da-1cca-4958-8838-4e7fc1364a78"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.741455 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "test-operator-logs") pod "11e9b8da-1cca-4958-8838-4e7fc1364a78" (UID: "11e9b8da-1cca-4958-8838-4e7fc1364a78"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.763432 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11e9b8da-1cca-4958-8838-4e7fc1364a78-tobiko-config" (OuterVolumeSpecName: "tobiko-config") pod "11e9b8da-1cca-4958-8838-4e7fc1364a78" (UID: "11e9b8da-1cca-4958-8838-4e7fc1364a78"). InnerVolumeSpecName "tobiko-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.763775 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11e9b8da-1cca-4958-8838-4e7fc1364a78-tobiko-private-key" (OuterVolumeSpecName: "tobiko-private-key") pod "11e9b8da-1cca-4958-8838-4e7fc1364a78" (UID: "11e9b8da-1cca-4958-8838-4e7fc1364a78"). InnerVolumeSpecName "tobiko-private-key". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.769744 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11e9b8da-1cca-4958-8838-4e7fc1364a78-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "11e9b8da-1cca-4958-8838-4e7fc1364a78" (UID: "11e9b8da-1cca-4958-8838-4e7fc1364a78"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.781342 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11e9b8da-1cca-4958-8838-4e7fc1364a78-kubeconfig" (OuterVolumeSpecName: "kubeconfig") pod "11e9b8da-1cca-4958-8838-4e7fc1364a78" (UID: "11e9b8da-1cca-4958-8838-4e7fc1364a78"). InnerVolumeSpecName "kubeconfig". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.785328 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11e9b8da-1cca-4958-8838-4e7fc1364a78-tobiko-public-key" (OuterVolumeSpecName: "tobiko-public-key") pod "11e9b8da-1cca-4958-8838-4e7fc1364a78" (UID: "11e9b8da-1cca-4958-8838-4e7fc1364a78"). InnerVolumeSpecName "tobiko-public-key". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.789997 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11e9b8da-1cca-4958-8838-4e7fc1364a78-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "11e9b8da-1cca-4958-8838-4e7fc1364a78" (UID: "11e9b8da-1cca-4958-8838-4e7fc1364a78"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.797749 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11e9b8da-1cca-4958-8838-4e7fc1364a78-test-operator-clouds-config" (OuterVolumeSpecName: "test-operator-clouds-config") pod "11e9b8da-1cca-4958-8838-4e7fc1364a78" (UID: "11e9b8da-1cca-4958-8838-4e7fc1364a78"). InnerVolumeSpecName "test-operator-clouds-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.837683 4959 reconciler_common.go:293] "Volume detached for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/11e9b8da-1cca-4958-8838-4e7fc1364a78-test-operator-clouds-config\") on node \"crc\" DevicePath \"\"" Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.837719 4959 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/11e9b8da-1cca-4958-8838-4e7fc1364a78-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.837757 4959 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.837772 4959 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/11e9b8da-1cca-4958-8838-4e7fc1364a78-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.837783 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/11e9b8da-1cca-4958-8838-4e7fc1364a78-ceph\") on node \"crc\" DevicePath \"\"" Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.837792 4959 reconciler_common.go:293] "Volume detached for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/11e9b8da-1cca-4958-8838-4e7fc1364a78-tobiko-public-key\") on node \"crc\" DevicePath \"\"" Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.837803 4959 reconciler_common.go:293] "Volume detached for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/11e9b8da-1cca-4958-8838-4e7fc1364a78-tobiko-config\") on node \"crc\" DevicePath \"\"" Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.837813 4959 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/11e9b8da-1cca-4958-8838-4e7fc1364a78-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.837824 4959 reconciler_common.go:293] "Volume detached for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/11e9b8da-1cca-4958-8838-4e7fc1364a78-tobiko-private-key\") on node \"crc\" DevicePath \"\"" Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.837834 4959 reconciler_common.go:293] "Volume detached for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/11e9b8da-1cca-4958-8838-4e7fc1364a78-kubeconfig\") on node \"crc\" DevicePath \"\"" Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.837841 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtft9\" (UniqueName: \"kubernetes.io/projected/11e9b8da-1cca-4958-8838-4e7fc1364a78-kube-api-access-rtft9\") on node \"crc\" DevicePath \"\"" Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.864410 4959 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 07 15:30:56 crc kubenswrapper[4959]: I1007 15:30:56.940597 4959 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 07 15:30:57 crc kubenswrapper[4959]: I1007 15:30:57.139882 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s01-sanity" event={"ID":"11e9b8da-1cca-4958-8838-4e7fc1364a78","Type":"ContainerDied","Data":"77e556549e4d10304225ba03fd689a4455445c0ba2aaf8248ee4afc1fcfdb148"} Oct 07 15:30:57 crc kubenswrapper[4959]: I1007 15:30:57.139932 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77e556549e4d10304225ba03fd689a4455445c0ba2aaf8248ee4afc1fcfdb148" Oct 07 15:30:57 crc kubenswrapper[4959]: I1007 15:30:57.140299 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 07 15:30:58 crc kubenswrapper[4959]: I1007 15:30:58.254490 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11e9b8da-1cca-4958-8838-4e7fc1364a78-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "11e9b8da-1cca-4958-8838-4e7fc1364a78" (UID: "11e9b8da-1cca-4958-8838-4e7fc1364a78"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:30:58 crc kubenswrapper[4959]: I1007 15:30:58.268073 4959 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/11e9b8da-1cca-4958-8838-4e7fc1364a78-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 07 15:31:09 crc kubenswrapper[4959]: I1007 15:31:09.422301 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko"] Oct 07 15:31:09 crc kubenswrapper[4959]: E1007 15:31:09.423293 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11e9b8da-1cca-4958-8838-4e7fc1364a78" containerName="tobiko-tests-tobiko" Oct 07 15:31:09 crc kubenswrapper[4959]: I1007 15:31:09.423311 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="11e9b8da-1cca-4958-8838-4e7fc1364a78" containerName="tobiko-tests-tobiko" Oct 07 15:31:09 crc kubenswrapper[4959]: E1007 15:31:09.423369 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df241fc3-4163-4d9a-9b82-af3bc490dd85" containerName="collect-profiles" Oct 07 15:31:09 crc kubenswrapper[4959]: I1007 15:31:09.423377 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="df241fc3-4163-4d9a-9b82-af3bc490dd85" containerName="collect-profiles" Oct 07 15:31:09 crc kubenswrapper[4959]: I1007 15:31:09.423616 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="11e9b8da-1cca-4958-8838-4e7fc1364a78" containerName="tobiko-tests-tobiko" Oct 07 15:31:09 crc kubenswrapper[4959]: I1007 15:31:09.423674 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="df241fc3-4163-4d9a-9b82-af3bc490dd85" containerName="collect-profiles" Oct 07 15:31:09 crc kubenswrapper[4959]: I1007 15:31:09.424522 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Oct 07 15:31:09 crc kubenswrapper[4959]: I1007 15:31:09.432747 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko"] Oct 07 15:31:09 crc kubenswrapper[4959]: I1007 15:31:09.511338 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc7gn\" (UniqueName: \"kubernetes.io/projected/e58ced67-303b-467a-b929-7de1e89d2317-kube-api-access-fc7gn\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"e58ced67-303b-467a-b929-7de1e89d2317\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Oct 07 15:31:09 crc kubenswrapper[4959]: I1007 15:31:09.511410 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"e58ced67-303b-467a-b929-7de1e89d2317\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Oct 07 15:31:09 crc kubenswrapper[4959]: I1007 15:31:09.613866 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc7gn\" (UniqueName: \"kubernetes.io/projected/e58ced67-303b-467a-b929-7de1e89d2317-kube-api-access-fc7gn\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"e58ced67-303b-467a-b929-7de1e89d2317\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Oct 07 15:31:09 crc kubenswrapper[4959]: I1007 15:31:09.613951 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"e58ced67-303b-467a-b929-7de1e89d2317\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Oct 07 15:31:09 crc kubenswrapper[4959]: I1007 15:31:09.614997 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"e58ced67-303b-467a-b929-7de1e89d2317\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Oct 07 15:31:09 crc kubenswrapper[4959]: I1007 15:31:09.640501 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc7gn\" (UniqueName: \"kubernetes.io/projected/e58ced67-303b-467a-b929-7de1e89d2317-kube-api-access-fc7gn\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"e58ced67-303b-467a-b929-7de1e89d2317\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Oct 07 15:31:09 crc kubenswrapper[4959]: I1007 15:31:09.656464 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"e58ced67-303b-467a-b929-7de1e89d2317\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Oct 07 15:31:09 crc kubenswrapper[4959]: I1007 15:31:09.788564 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Oct 07 15:31:10 crc kubenswrapper[4959]: I1007 15:31:10.250013 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 15:31:10 crc kubenswrapper[4959]: I1007 15:31:10.250159 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko"] Oct 07 15:31:11 crc kubenswrapper[4959]: I1007 15:31:11.265088 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" event={"ID":"e58ced67-303b-467a-b929-7de1e89d2317","Type":"ContainerStarted","Data":"d8de79b470e11e3c7411a858e0f493d202f9c92f7db4e9f21dfde33ef46dc159"} Oct 07 15:31:11 crc kubenswrapper[4959]: I1007 15:31:11.265898 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" event={"ID":"e58ced67-303b-467a-b929-7de1e89d2317","Type":"ContainerStarted","Data":"7dfbbf3fa4d5c3a65dfd81bd57716de230818bed6c284c0bb4b01f9704d8a9b9"} Oct 07 15:31:11 crc kubenswrapper[4959]: I1007 15:31:11.285623 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" podStartSLOduration=1.607447597 podStartE2EDuration="2.285602605s" podCreationTimestamp="2025-10-07 15:31:09 +0000 UTC" firstStartedPulling="2025-10-07 15:31:10.249746598 +0000 UTC m=+6352.333150913" lastFinishedPulling="2025-10-07 15:31:10.927901596 +0000 UTC m=+6353.011305921" observedRunningTime="2025-10-07 15:31:11.281718426 +0000 UTC m=+6353.365122751" watchObservedRunningTime="2025-10-07 15:31:11.285602605 +0000 UTC m=+6353.369006930" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.013742 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ansibletest-ansibletest"] Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.015708 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.018910 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.019301 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.034624 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ansibletest-ansibletest"] Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.053802 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-ca-certs\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.053882 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-openstack-config-secret\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.053979 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-compute-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.054002 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-test-operator-ephemeral-temporary\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.054020 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44k7x\" (UniqueName: \"kubernetes.io/projected/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-kube-api-access-44k7x\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.054054 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.054074 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-openstack-config\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.054121 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-test-operator-ephemeral-workdir\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.054152 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-ceph\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.054193 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-workload-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.155439 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-test-operator-ephemeral-workdir\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.155762 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-ceph\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.155882 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-workload-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.156030 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-ca-certs\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.156152 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-test-operator-ephemeral-workdir\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.156315 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-openstack-config-secret\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.156509 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-compute-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.156612 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-test-operator-ephemeral-temporary\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.156789 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44k7x\" (UniqueName: \"kubernetes.io/projected/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-kube-api-access-44k7x\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.156956 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.157087 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-openstack-config\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.157215 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.157527 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-test-operator-ephemeral-temporary\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.158572 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-openstack-config\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.164158 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-workload-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.165545 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-compute-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.167827 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-openstack-config-secret\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.173852 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-ca-certs\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.174927 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44k7x\" (UniqueName: \"kubernetes.io/projected/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-kube-api-access-44k7x\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.176725 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-ceph\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.195603 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ansibletest-ansibletest\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.347384 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ansibletest-ansibletest" Oct 07 15:31:30 crc kubenswrapper[4959]: I1007 15:31:30.813432 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ansibletest-ansibletest"] Oct 07 15:31:30 crc kubenswrapper[4959]: W1007 15:31:30.824739 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod954ed5c7_ad7c_4cd0_ac69_ffa0c9fa3c34.slice/crio-8bf423d052313ad5287aa5e69739a3bc30bb913b761ff3daf29b02a41c8db751 WatchSource:0}: Error finding container 8bf423d052313ad5287aa5e69739a3bc30bb913b761ff3daf29b02a41c8db751: Status 404 returned error can't find the container with id 8bf423d052313ad5287aa5e69739a3bc30bb913b761ff3daf29b02a41c8db751 Oct 07 15:31:31 crc kubenswrapper[4959]: I1007 15:31:31.467114 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ansibletest-ansibletest" event={"ID":"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34","Type":"ContainerStarted","Data":"8bf423d052313ad5287aa5e69739a3bc30bb913b761ff3daf29b02a41c8db751"} Oct 07 15:31:44 crc kubenswrapper[4959]: E1007 15:31:44.438166 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ansible-tests:current-podified" Oct 07 15:31:44 crc kubenswrapper[4959]: E1007 15:31:44.440343 4959 kuberuntime_manager.go:1274] "Unhandled Error" err=< Oct 07 15:31:44 crc kubenswrapper[4959]: container &Container{Name:ansibletest-ansibletest,Image:quay.io/podified-antelope-centos9/openstack-ansible-tests:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:POD_ANSIBLE_EXTRA_VARS,Value:-e manual_run=false,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_FILE_EXTRA_VARS,Value:--- Oct 07 15:31:44 crc kubenswrapper[4959]: foo: bar Oct 07 15:31:44 crc kubenswrapper[4959]: ,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_GIT_BRANCH,Value:,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_GIT_REPO,Value:https://github.com/ansible/test-playbooks,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_INVENTORY,Value:localhost ansible_connection=local ansible_python_interpreter=python3 Oct 07 15:31:44 crc kubenswrapper[4959]: ,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_PLAYBOOK,Value:./debug.yml,ValueFrom:nil,},EnvVar{Name:POD_DEBUG,Value:false,ValueFrom:nil,},EnvVar{Name:POD_INSTALL_COLLECTIONS,Value:,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{4 0} {} 4 DecimalSI},memory: {{4294967296 0} {} 4Gi BinarySI},},Requests:ResourceList{cpu: {{2 0} {} 2 DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/ansible,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/AnsibleTests/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/ansible/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/var/lib/ansible/.config/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ca-bundle.trust.crt,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:workload-ssh-secret,ReadOnly:true,MountPath:/var/lib/ansible/test_keypair.key,SubPath:ssh-privatekey,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:compute-ssh-secret,ReadOnly:true,MountPath:/var/lib/ansible/.ssh/compute_id,SubPath:ssh-privatekey,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceph,ReadOnly:true,MountPath:/etc/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-44k7x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN NET_RAW],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*227,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*227,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ansibletest-ansibletest_openstack(954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Oct 07 15:31:44 crc kubenswrapper[4959]: > logger="UnhandledError" Oct 07 15:31:44 crc kubenswrapper[4959]: E1007 15:31:44.441867 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ansibletest-ansibletest\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ansibletest-ansibletest" podUID="954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34" Oct 07 15:31:44 crc kubenswrapper[4959]: E1007 15:31:44.584482 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ansibletest-ansibletest\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ansible-tests:current-podified\\\"\"" pod="openstack/ansibletest-ansibletest" podUID="954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34" Oct 07 15:31:50 crc kubenswrapper[4959]: I1007 15:31:50.631847 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 15:31:50 crc kubenswrapper[4959]: I1007 15:31:50.632482 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 15:32:00 crc kubenswrapper[4959]: I1007 15:32:00.716729 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ansibletest-ansibletest" event={"ID":"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34","Type":"ContainerStarted","Data":"aaf56d5ecd31ef91fdba0f92de68c1655440ddb63966eb637a13d4afd96abeef"} Oct 07 15:32:00 crc kubenswrapper[4959]: I1007 15:32:00.734939 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ansibletest-ansibletest" podStartSLOduration=4.045295787 podStartE2EDuration="32.734916625s" podCreationTimestamp="2025-10-07 15:31:28 +0000 UTC" firstStartedPulling="2025-10-07 15:31:30.827001547 +0000 UTC m=+6372.910405872" lastFinishedPulling="2025-10-07 15:31:59.516622375 +0000 UTC m=+6401.600026710" observedRunningTime="2025-10-07 15:32:00.732476713 +0000 UTC m=+6402.815881048" watchObservedRunningTime="2025-10-07 15:32:00.734916625 +0000 UTC m=+6402.818320950" Oct 07 15:32:02 crc kubenswrapper[4959]: I1007 15:32:02.739910 4959 generic.go:334] "Generic (PLEG): container finished" podID="954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34" containerID="aaf56d5ecd31ef91fdba0f92de68c1655440ddb63966eb637a13d4afd96abeef" exitCode=0 Oct 07 15:32:02 crc kubenswrapper[4959]: I1007 15:32:02.739981 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ansibletest-ansibletest" event={"ID":"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34","Type":"ContainerDied","Data":"aaf56d5ecd31ef91fdba0f92de68c1655440ddb63966eb637a13d4afd96abeef"} Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.063940 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ansibletest-ansibletest" Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.219178 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-ceph\") pod \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.219243 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44k7x\" (UniqueName: \"kubernetes.io/projected/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-kube-api-access-44k7x\") pod \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.219270 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-ca-certs\") pod \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.219300 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.219363 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-test-operator-ephemeral-temporary\") pod \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.219445 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-test-operator-ephemeral-workdir\") pod \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.219502 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-compute-ssh-secret\") pod \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.219528 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-workload-ssh-secret\") pod \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.219571 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-openstack-config\") pod \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.219592 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-openstack-config-secret\") pod \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\" (UID: \"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34\") " Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.220079 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34" (UID: "954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.226130 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "test-operator-logs") pod "954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34" (UID: "954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.230864 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-kube-api-access-44k7x" (OuterVolumeSpecName: "kube-api-access-44k7x") pod "954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34" (UID: "954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34"). InnerVolumeSpecName "kube-api-access-44k7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.231020 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-ceph" (OuterVolumeSpecName: "ceph") pod "954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34" (UID: "954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.236477 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34" (UID: "954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.252336 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34" (UID: "954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.253509 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-compute-ssh-secret" (OuterVolumeSpecName: "compute-ssh-secret") pod "954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34" (UID: "954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34"). InnerVolumeSpecName "compute-ssh-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.264358 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-workload-ssh-secret" (OuterVolumeSpecName: "workload-ssh-secret") pod "954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34" (UID: "954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34"). InnerVolumeSpecName "workload-ssh-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.282559 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34" (UID: "954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.290935 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34" (UID: "954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.323122 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-ceph\") on node \"crc\" DevicePath \"\"" Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.323154 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44k7x\" (UniqueName: \"kubernetes.io/projected/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-kube-api-access-44k7x\") on node \"crc\" DevicePath \"\"" Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.323167 4959 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.323214 4959 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.323229 4959 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.323241 4959 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.323253 4959 reconciler_common.go:293] "Volume detached for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-compute-ssh-secret\") on node \"crc\" DevicePath \"\"" Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.323262 4959 reconciler_common.go:293] "Volume detached for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-workload-ssh-secret\") on node \"crc\" DevicePath \"\"" Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.323270 4959 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.323278 4959 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.347018 4959 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.426065 4959 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.767012 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ansibletest-ansibletest" event={"ID":"954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34","Type":"ContainerDied","Data":"8bf423d052313ad5287aa5e69739a3bc30bb913b761ff3daf29b02a41c8db751"} Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.767078 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8bf423d052313ad5287aa5e69739a3bc30bb913b761ff3daf29b02a41c8db751" Oct 07 15:32:04 crc kubenswrapper[4959]: I1007 15:32:04.767202 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ansibletest-ansibletest" Oct 07 15:32:15 crc kubenswrapper[4959]: I1007 15:32:15.219929 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest"] Oct 07 15:32:15 crc kubenswrapper[4959]: E1007 15:32:15.221036 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34" containerName="ansibletest-ansibletest" Oct 07 15:32:15 crc kubenswrapper[4959]: I1007 15:32:15.221058 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34" containerName="ansibletest-ansibletest" Oct 07 15:32:15 crc kubenswrapper[4959]: I1007 15:32:15.221331 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34" containerName="ansibletest-ansibletest" Oct 07 15:32:15 crc kubenswrapper[4959]: I1007 15:32:15.222513 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Oct 07 15:32:15 crc kubenswrapper[4959]: I1007 15:32:15.241303 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest"] Oct 07 15:32:15 crc kubenswrapper[4959]: I1007 15:32:15.364916 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"409bd790-46de-4dc6-bfae-6a75d5117a74\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Oct 07 15:32:15 crc kubenswrapper[4959]: I1007 15:32:15.365339 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7dv6\" (UniqueName: \"kubernetes.io/projected/409bd790-46de-4dc6-bfae-6a75d5117a74-kube-api-access-v7dv6\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"409bd790-46de-4dc6-bfae-6a75d5117a74\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Oct 07 15:32:15 crc kubenswrapper[4959]: I1007 15:32:15.468081 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"409bd790-46de-4dc6-bfae-6a75d5117a74\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Oct 07 15:32:15 crc kubenswrapper[4959]: I1007 15:32:15.468199 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7dv6\" (UniqueName: \"kubernetes.io/projected/409bd790-46de-4dc6-bfae-6a75d5117a74-kube-api-access-v7dv6\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"409bd790-46de-4dc6-bfae-6a75d5117a74\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Oct 07 15:32:15 crc kubenswrapper[4959]: I1007 15:32:15.468883 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"409bd790-46de-4dc6-bfae-6a75d5117a74\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Oct 07 15:32:15 crc kubenswrapper[4959]: I1007 15:32:15.493450 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7dv6\" (UniqueName: \"kubernetes.io/projected/409bd790-46de-4dc6-bfae-6a75d5117a74-kube-api-access-v7dv6\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"409bd790-46de-4dc6-bfae-6a75d5117a74\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Oct 07 15:32:15 crc kubenswrapper[4959]: I1007 15:32:15.516254 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"409bd790-46de-4dc6-bfae-6a75d5117a74\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Oct 07 15:32:15 crc kubenswrapper[4959]: I1007 15:32:15.552521 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Oct 07 15:32:16 crc kubenswrapper[4959]: I1007 15:32:16.083130 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest"] Oct 07 15:32:16 crc kubenswrapper[4959]: I1007 15:32:16.901656 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" event={"ID":"409bd790-46de-4dc6-bfae-6a75d5117a74","Type":"ContainerStarted","Data":"6e8881d8587a9be3955ced60ec68da723d71a6544e9aca87f546e51b47639d60"} Oct 07 15:32:16 crc kubenswrapper[4959]: I1007 15:32:16.902139 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" event={"ID":"409bd790-46de-4dc6-bfae-6a75d5117a74","Type":"ContainerStarted","Data":"e38757c854a028f3944f1874b15f40d59de9494799749f5f328966797a6d4595"} Oct 07 15:32:16 crc kubenswrapper[4959]: I1007 15:32:16.927166 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" podStartSLOduration=1.4010431190000001 podStartE2EDuration="1.927122831s" podCreationTimestamp="2025-10-07 15:32:15 +0000 UTC" firstStartedPulling="2025-10-07 15:32:16.087642199 +0000 UTC m=+6418.171046524" lastFinishedPulling="2025-10-07 15:32:16.613721901 +0000 UTC m=+6418.697126236" observedRunningTime="2025-10-07 15:32:16.916672335 +0000 UTC m=+6419.000076680" watchObservedRunningTime="2025-10-07 15:32:16.927122831 +0000 UTC m=+6419.010527176" Oct 07 15:32:20 crc kubenswrapper[4959]: I1007 15:32:20.629668 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 15:32:20 crc kubenswrapper[4959]: I1007 15:32:20.630315 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.579837 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizontest-tests-horizontest"] Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.581811 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizontest-tests-horizontest" Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.584815 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizontest-tests-horizontesthorizontest-config" Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.585414 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"test-operator-clouds-config" Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.596164 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizontest-tests-horizontest"] Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.703246 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8ae1c80c-a359-4751-80e1-58da5918db21-ca-certs\") pod \"horizontest-tests-horizontest\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " pod="openstack/horizontest-tests-horizontest" Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.703564 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgjhz\" (UniqueName: \"kubernetes.io/projected/8ae1c80c-a359-4751-80e1-58da5918db21-kube-api-access-bgjhz\") pod \"horizontest-tests-horizontest\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " pod="openstack/horizontest-tests-horizontest" Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.703621 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/8ae1c80c-a359-4751-80e1-58da5918db21-test-operator-clouds-config\") pod \"horizontest-tests-horizontest\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " pod="openstack/horizontest-tests-horizontest" Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.703652 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"horizontest-tests-horizontest\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " pod="openstack/horizontest-tests-horizontest" Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.703674 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8ae1c80c-a359-4751-80e1-58da5918db21-ceph\") pod \"horizontest-tests-horizontest\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " pod="openstack/horizontest-tests-horizontest" Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.703875 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8ae1c80c-a359-4751-80e1-58da5918db21-test-operator-ephemeral-temporary\") pod \"horizontest-tests-horizontest\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " pod="openstack/horizontest-tests-horizontest" Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.703940 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8ae1c80c-a359-4751-80e1-58da5918db21-test-operator-ephemeral-workdir\") pod \"horizontest-tests-horizontest\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " pod="openstack/horizontest-tests-horizontest" Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.704215 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8ae1c80c-a359-4751-80e1-58da5918db21-openstack-config-secret\") pod \"horizontest-tests-horizontest\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " pod="openstack/horizontest-tests-horizontest" Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.805687 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/8ae1c80c-a359-4751-80e1-58da5918db21-test-operator-clouds-config\") pod \"horizontest-tests-horizontest\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " pod="openstack/horizontest-tests-horizontest" Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.805771 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"horizontest-tests-horizontest\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " pod="openstack/horizontest-tests-horizontest" Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.805803 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8ae1c80c-a359-4751-80e1-58da5918db21-ceph\") pod \"horizontest-tests-horizontest\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " pod="openstack/horizontest-tests-horizontest" Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.805887 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8ae1c80c-a359-4751-80e1-58da5918db21-test-operator-ephemeral-temporary\") pod \"horizontest-tests-horizontest\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " pod="openstack/horizontest-tests-horizontest" Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.805925 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8ae1c80c-a359-4751-80e1-58da5918db21-test-operator-ephemeral-workdir\") pod \"horizontest-tests-horizontest\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " pod="openstack/horizontest-tests-horizontest" Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.806128 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8ae1c80c-a359-4751-80e1-58da5918db21-openstack-config-secret\") pod \"horizontest-tests-horizontest\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " pod="openstack/horizontest-tests-horizontest" Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.806204 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8ae1c80c-a359-4751-80e1-58da5918db21-ca-certs\") pod \"horizontest-tests-horizontest\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " pod="openstack/horizontest-tests-horizontest" Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.806242 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgjhz\" (UniqueName: \"kubernetes.io/projected/8ae1c80c-a359-4751-80e1-58da5918db21-kube-api-access-bgjhz\") pod \"horizontest-tests-horizontest\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " pod="openstack/horizontest-tests-horizontest" Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.807263 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8ae1c80c-a359-4751-80e1-58da5918db21-test-operator-ephemeral-temporary\") pod \"horizontest-tests-horizontest\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " pod="openstack/horizontest-tests-horizontest" Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.807261 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/8ae1c80c-a359-4751-80e1-58da5918db21-test-operator-clouds-config\") pod \"horizontest-tests-horizontest\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " pod="openstack/horizontest-tests-horizontest" Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.807817 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"horizontest-tests-horizontest\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/horizontest-tests-horizontest" Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.807887 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8ae1c80c-a359-4751-80e1-58da5918db21-test-operator-ephemeral-workdir\") pod \"horizontest-tests-horizontest\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " pod="openstack/horizontest-tests-horizontest" Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.813432 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8ae1c80c-a359-4751-80e1-58da5918db21-openstack-config-secret\") pod \"horizontest-tests-horizontest\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " pod="openstack/horizontest-tests-horizontest" Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.813653 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8ae1c80c-a359-4751-80e1-58da5918db21-ceph\") pod \"horizontest-tests-horizontest\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " pod="openstack/horizontest-tests-horizontest" Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.813756 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8ae1c80c-a359-4751-80e1-58da5918db21-ca-certs\") pod \"horizontest-tests-horizontest\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " pod="openstack/horizontest-tests-horizontest" Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.829913 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgjhz\" (UniqueName: \"kubernetes.io/projected/8ae1c80c-a359-4751-80e1-58da5918db21-kube-api-access-bgjhz\") pod \"horizontest-tests-horizontest\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " pod="openstack/horizontest-tests-horizontest" Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.836789 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"horizontest-tests-horizontest\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " pod="openstack/horizontest-tests-horizontest" Oct 07 15:32:35 crc kubenswrapper[4959]: I1007 15:32:35.909328 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizontest-tests-horizontest" Oct 07 15:32:36 crc kubenswrapper[4959]: I1007 15:32:36.337534 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizontest-tests-horizontest"] Oct 07 15:32:37 crc kubenswrapper[4959]: I1007 15:32:37.082435 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizontest-tests-horizontest" event={"ID":"8ae1c80c-a359-4751-80e1-58da5918db21","Type":"ContainerStarted","Data":"fcdd85214ccf6095fe5d4dbae9b62293968f45135475017a2ef7973eb91c923d"} Oct 07 15:32:50 crc kubenswrapper[4959]: I1007 15:32:50.629735 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 15:32:50 crc kubenswrapper[4959]: I1007 15:32:50.630328 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 15:32:50 crc kubenswrapper[4959]: I1007 15:32:50.630378 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 15:32:50 crc kubenswrapper[4959]: I1007 15:32:50.631204 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a"} pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 15:32:50 crc kubenswrapper[4959]: I1007 15:32:50.631254 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" containerID="cri-o://02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a" gracePeriod=600 Oct 07 15:32:52 crc kubenswrapper[4959]: I1007 15:32:52.245627 4959 generic.go:334] "Generic (PLEG): container finished" podID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerID="02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a" exitCode=0 Oct 07 15:32:52 crc kubenswrapper[4959]: I1007 15:32:52.246063 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerDied","Data":"02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a"} Oct 07 15:32:52 crc kubenswrapper[4959]: I1007 15:32:52.246133 4959 scope.go:117] "RemoveContainer" containerID="3b696f2313816349c23c9e18632aa1ab844c286a6de0e092da7dc16410d9a66e" Oct 07 15:32:54 crc kubenswrapper[4959]: E1007 15:32:54.487465 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:32:54 crc kubenswrapper[4959]: E1007 15:32:54.571664 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizontest:current-podified" Oct 07 15:32:54 crc kubenswrapper[4959]: E1007 15:32:54.572371 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizontest-tests-horizontest,Image:quay.io/podified-antelope-centos9/openstack-horizontest:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADMIN_PASSWORD,Value:12345678,ValueFrom:nil,},EnvVar{Name:ADMIN_USERNAME,Value:admin,ValueFrom:nil,},EnvVar{Name:AUTH_URL,Value:https://keystone-public-openstack.apps-crc.testing,ValueFrom:nil,},EnvVar{Name:DASHBOARD_URL,Value:https://horizon-openstack.apps-crc.testing/,ValueFrom:nil,},EnvVar{Name:EXTRA_FLAG,Value:not pagination and test_users.py,ValueFrom:nil,},EnvVar{Name:FLAVOR_NAME,Value:m1.tiny,ValueFrom:nil,},EnvVar{Name:HORIZONTEST_DEBUG_MODE,Value:false,ValueFrom:nil,},EnvVar{Name:HORIZON_KEYS_FOLDER,Value:/etc/test_operator,ValueFrom:nil,},EnvVar{Name:HORIZON_LOGS_DIR_NAME,Value:horizon,ValueFrom:nil,},EnvVar{Name:HORIZON_REPO_BRANCH,Value:master,ValueFrom:nil,},EnvVar{Name:IMAGE_FILE,Value:/var/lib/horizontest/cirros-0.6.2-x86_64-disk.img,ValueFrom:nil,},EnvVar{Name:IMAGE_FILE_NAME,Value:cirros-0.6.2-x86_64-disk,ValueFrom:nil,},EnvVar{Name:IMAGE_URL,Value:http://download.cirros-cloud.net/0.6.2/cirros-0.6.2-x86_64-disk.img,ValueFrom:nil,},EnvVar{Name:PASSWORD,Value:horizontest,ValueFrom:nil,},EnvVar{Name:PROJECT_NAME,Value:horizontest,ValueFrom:nil,},EnvVar{Name:PROJECT_NAME_XPATH,Value://*[@class=\"context-project\"]//ancestor::ul,ValueFrom:nil,},EnvVar{Name:REPO_URL,Value:https://review.opendev.org/openstack/horizon,ValueFrom:nil,},EnvVar{Name:USER_NAME,Value:horizontest,ValueFrom:nil,},EnvVar{Name:USE_EXTERNAL_FILES,Value:True,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{2 0} {} 2 DecimalSI},memory: {{4294967296 0} {} 4Gi BinarySI},},Requests:ResourceList{cpu: {{1 0} {} 1 DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/horizontest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/horizontest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-clouds-config,ReadOnly:true,MountPath:/var/lib/horizontest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-clouds-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ca-bundle.trust.crt,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceph,ReadOnly:true,MountPath:/etc/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bgjhz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN NET_RAW],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42455,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42455,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizontest-tests-horizontest_openstack(8ae1c80c-a359-4751-80e1-58da5918db21): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 07 15:32:54 crc kubenswrapper[4959]: E1007 15:32:54.573686 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"horizontest-tests-horizontest\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/horizontest-tests-horizontest" podUID="8ae1c80c-a359-4751-80e1-58da5918db21" Oct 07 15:32:55 crc kubenswrapper[4959]: I1007 15:32:55.274329 4959 scope.go:117] "RemoveContainer" containerID="02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a" Oct 07 15:32:55 crc kubenswrapper[4959]: E1007 15:32:55.275063 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"horizontest-tests-horizontest\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizontest:current-podified\\\"\"" pod="openstack/horizontest-tests-horizontest" podUID="8ae1c80c-a359-4751-80e1-58da5918db21" Oct 07 15:32:55 crc kubenswrapper[4959]: E1007 15:32:55.275250 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:33:07 crc kubenswrapper[4959]: I1007 15:33:07.653945 4959 scope.go:117] "RemoveContainer" containerID="02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a" Oct 07 15:33:07 crc kubenswrapper[4959]: E1007 15:33:07.654736 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:33:10 crc kubenswrapper[4959]: I1007 15:33:10.457161 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizontest-tests-horizontest" event={"ID":"8ae1c80c-a359-4751-80e1-58da5918db21","Type":"ContainerStarted","Data":"a5c1a57d45218c7c64404951e8402885c7af6b4fd8dd5cc0609124351fca6063"} Oct 07 15:33:10 crc kubenswrapper[4959]: I1007 15:33:10.488168 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizontest-tests-horizontest" podStartSLOduration=3.977919108 podStartE2EDuration="36.488143878s" podCreationTimestamp="2025-10-07 15:32:34 +0000 UTC" firstStartedPulling="2025-10-07 15:32:36.344819362 +0000 UTC m=+6438.428223687" lastFinishedPulling="2025-10-07 15:33:08.855044142 +0000 UTC m=+6470.938448457" observedRunningTime="2025-10-07 15:33:10.480134863 +0000 UTC m=+6472.563539188" watchObservedRunningTime="2025-10-07 15:33:10.488143878 +0000 UTC m=+6472.571548203" Oct 07 15:33:18 crc kubenswrapper[4959]: I1007 15:33:18.663272 4959 scope.go:117] "RemoveContainer" containerID="02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a" Oct 07 15:33:18 crc kubenswrapper[4959]: E1007 15:33:18.664313 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:33:33 crc kubenswrapper[4959]: I1007 15:33:33.654518 4959 scope.go:117] "RemoveContainer" containerID="02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a" Oct 07 15:33:33 crc kubenswrapper[4959]: E1007 15:33:33.655431 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:33:48 crc kubenswrapper[4959]: I1007 15:33:48.661310 4959 scope.go:117] "RemoveContainer" containerID="02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a" Oct 07 15:33:48 crc kubenswrapper[4959]: E1007 15:33:48.663418 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:34:03 crc kubenswrapper[4959]: I1007 15:34:03.654213 4959 scope.go:117] "RemoveContainer" containerID="02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a" Oct 07 15:34:03 crc kubenswrapper[4959]: E1007 15:34:03.655164 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:34:17 crc kubenswrapper[4959]: I1007 15:34:17.653344 4959 scope.go:117] "RemoveContainer" containerID="02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a" Oct 07 15:34:17 crc kubenswrapper[4959]: E1007 15:34:17.654230 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:34:31 crc kubenswrapper[4959]: I1007 15:34:31.654404 4959 scope.go:117] "RemoveContainer" containerID="02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a" Oct 07 15:34:31 crc kubenswrapper[4959]: E1007 15:34:31.655155 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:34:43 crc kubenswrapper[4959]: I1007 15:34:43.653393 4959 scope.go:117] "RemoveContainer" containerID="02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a" Oct 07 15:34:43 crc kubenswrapper[4959]: E1007 15:34:43.654089 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:34:56 crc kubenswrapper[4959]: I1007 15:34:56.654163 4959 scope.go:117] "RemoveContainer" containerID="02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a" Oct 07 15:34:56 crc kubenswrapper[4959]: E1007 15:34:56.654977 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:35:03 crc kubenswrapper[4959]: I1007 15:35:03.506172 4959 generic.go:334] "Generic (PLEG): container finished" podID="8ae1c80c-a359-4751-80e1-58da5918db21" containerID="a5c1a57d45218c7c64404951e8402885c7af6b4fd8dd5cc0609124351fca6063" exitCode=0 Oct 07 15:35:03 crc kubenswrapper[4959]: I1007 15:35:03.506263 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizontest-tests-horizontest" event={"ID":"8ae1c80c-a359-4751-80e1-58da5918db21","Type":"ContainerDied","Data":"a5c1a57d45218c7c64404951e8402885c7af6b4fd8dd5cc0609124351fca6063"} Oct 07 15:35:04 crc kubenswrapper[4959]: I1007 15:35:04.854863 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizontest-tests-horizontest" Oct 07 15:35:05 crc kubenswrapper[4959]: I1007 15:35:05.026091 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8ae1c80c-a359-4751-80e1-58da5918db21-test-operator-ephemeral-temporary\") pod \"8ae1c80c-a359-4751-80e1-58da5918db21\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " Oct 07 15:35:05 crc kubenswrapper[4959]: I1007 15:35:05.026503 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8ae1c80c-a359-4751-80e1-58da5918db21-test-operator-ephemeral-workdir\") pod \"8ae1c80c-a359-4751-80e1-58da5918db21\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " Oct 07 15:35:05 crc kubenswrapper[4959]: I1007 15:35:05.026656 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8ae1c80c-a359-4751-80e1-58da5918db21-openstack-config-secret\") pod \"8ae1c80c-a359-4751-80e1-58da5918db21\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " Oct 07 15:35:05 crc kubenswrapper[4959]: I1007 15:35:05.026777 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgjhz\" (UniqueName: \"kubernetes.io/projected/8ae1c80c-a359-4751-80e1-58da5918db21-kube-api-access-bgjhz\") pod \"8ae1c80c-a359-4751-80e1-58da5918db21\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " Oct 07 15:35:05 crc kubenswrapper[4959]: I1007 15:35:05.026873 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8ae1c80c-a359-4751-80e1-58da5918db21-ca-certs\") pod \"8ae1c80c-a359-4751-80e1-58da5918db21\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " Oct 07 15:35:05 crc kubenswrapper[4959]: I1007 15:35:05.026972 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/8ae1c80c-a359-4751-80e1-58da5918db21-test-operator-clouds-config\") pod \"8ae1c80c-a359-4751-80e1-58da5918db21\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " Oct 07 15:35:05 crc kubenswrapper[4959]: I1007 15:35:05.027122 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8ae1c80c-a359-4751-80e1-58da5918db21-ceph\") pod \"8ae1c80c-a359-4751-80e1-58da5918db21\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " Oct 07 15:35:05 crc kubenswrapper[4959]: I1007 15:35:05.027253 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"8ae1c80c-a359-4751-80e1-58da5918db21\" (UID: \"8ae1c80c-a359-4751-80e1-58da5918db21\") " Oct 07 15:35:05 crc kubenswrapper[4959]: I1007 15:35:05.026885 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ae1c80c-a359-4751-80e1-58da5918db21-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "8ae1c80c-a359-4751-80e1-58da5918db21" (UID: "8ae1c80c-a359-4751-80e1-58da5918db21"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:35:05 crc kubenswrapper[4959]: I1007 15:35:05.028118 4959 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8ae1c80c-a359-4751-80e1-58da5918db21-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 07 15:35:05 crc kubenswrapper[4959]: I1007 15:35:05.032564 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "test-operator-logs") pod "8ae1c80c-a359-4751-80e1-58da5918db21" (UID: "8ae1c80c-a359-4751-80e1-58da5918db21"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 07 15:35:05 crc kubenswrapper[4959]: I1007 15:35:05.033049 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ae1c80c-a359-4751-80e1-58da5918db21-ceph" (OuterVolumeSpecName: "ceph") pod "8ae1c80c-a359-4751-80e1-58da5918db21" (UID: "8ae1c80c-a359-4751-80e1-58da5918db21"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:35:05 crc kubenswrapper[4959]: I1007 15:35:05.033193 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ae1c80c-a359-4751-80e1-58da5918db21-kube-api-access-bgjhz" (OuterVolumeSpecName: "kube-api-access-bgjhz") pod "8ae1c80c-a359-4751-80e1-58da5918db21" (UID: "8ae1c80c-a359-4751-80e1-58da5918db21"). InnerVolumeSpecName "kube-api-access-bgjhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 15:35:05 crc kubenswrapper[4959]: I1007 15:35:05.055587 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ae1c80c-a359-4751-80e1-58da5918db21-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "8ae1c80c-a359-4751-80e1-58da5918db21" (UID: "8ae1c80c-a359-4751-80e1-58da5918db21"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:35:05 crc kubenswrapper[4959]: I1007 15:35:05.077966 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ae1c80c-a359-4751-80e1-58da5918db21-test-operator-clouds-config" (OuterVolumeSpecName: "test-operator-clouds-config") pod "8ae1c80c-a359-4751-80e1-58da5918db21" (UID: "8ae1c80c-a359-4751-80e1-58da5918db21"). InnerVolumeSpecName "test-operator-clouds-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 15:35:05 crc kubenswrapper[4959]: I1007 15:35:05.088075 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ae1c80c-a359-4751-80e1-58da5918db21-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "8ae1c80c-a359-4751-80e1-58da5918db21" (UID: "8ae1c80c-a359-4751-80e1-58da5918db21"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:35:05 crc kubenswrapper[4959]: I1007 15:35:05.129601 4959 reconciler_common.go:293] "Volume detached for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/8ae1c80c-a359-4751-80e1-58da5918db21-test-operator-clouds-config\") on node \"crc\" DevicePath \"\"" Oct 07 15:35:05 crc kubenswrapper[4959]: I1007 15:35:05.129654 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8ae1c80c-a359-4751-80e1-58da5918db21-ceph\") on node \"crc\" DevicePath \"\"" Oct 07 15:35:05 crc kubenswrapper[4959]: I1007 15:35:05.129693 4959 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 07 15:35:05 crc kubenswrapper[4959]: I1007 15:35:05.129709 4959 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8ae1c80c-a359-4751-80e1-58da5918db21-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 07 15:35:05 crc kubenswrapper[4959]: I1007 15:35:05.129720 4959 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8ae1c80c-a359-4751-80e1-58da5918db21-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 07 15:35:05 crc kubenswrapper[4959]: I1007 15:35:05.129741 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgjhz\" (UniqueName: \"kubernetes.io/projected/8ae1c80c-a359-4751-80e1-58da5918db21-kube-api-access-bgjhz\") on node \"crc\" DevicePath \"\"" Oct 07 15:35:05 crc kubenswrapper[4959]: I1007 15:35:05.155624 4959 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 07 15:35:05 crc kubenswrapper[4959]: I1007 15:35:05.232390 4959 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 07 15:35:05 crc kubenswrapper[4959]: I1007 15:35:05.257299 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ae1c80c-a359-4751-80e1-58da5918db21-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "8ae1c80c-a359-4751-80e1-58da5918db21" (UID: "8ae1c80c-a359-4751-80e1-58da5918db21"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:35:05 crc kubenswrapper[4959]: I1007 15:35:05.333873 4959 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8ae1c80c-a359-4751-80e1-58da5918db21-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 07 15:35:05 crc kubenswrapper[4959]: I1007 15:35:05.526259 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizontest-tests-horizontest" event={"ID":"8ae1c80c-a359-4751-80e1-58da5918db21","Type":"ContainerDied","Data":"fcdd85214ccf6095fe5d4dbae9b62293968f45135475017a2ef7973eb91c923d"} Oct 07 15:35:05 crc kubenswrapper[4959]: I1007 15:35:05.526301 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fcdd85214ccf6095fe5d4dbae9b62293968f45135475017a2ef7973eb91c923d" Oct 07 15:35:05 crc kubenswrapper[4959]: I1007 15:35:05.526624 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizontest-tests-horizontest" Oct 07 15:35:08 crc kubenswrapper[4959]: I1007 15:35:08.663164 4959 scope.go:117] "RemoveContainer" containerID="02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a" Oct 07 15:35:08 crc kubenswrapper[4959]: E1007 15:35:08.664886 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:35:08 crc kubenswrapper[4959]: I1007 15:35:08.837063 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest"] Oct 07 15:35:08 crc kubenswrapper[4959]: E1007 15:35:08.837616 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ae1c80c-a359-4751-80e1-58da5918db21" containerName="horizontest-tests-horizontest" Oct 07 15:35:08 crc kubenswrapper[4959]: I1007 15:35:08.837636 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ae1c80c-a359-4751-80e1-58da5918db21" containerName="horizontest-tests-horizontest" Oct 07 15:35:08 crc kubenswrapper[4959]: I1007 15:35:08.837833 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ae1c80c-a359-4751-80e1-58da5918db21" containerName="horizontest-tests-horizontest" Oct 07 15:35:08 crc kubenswrapper[4959]: I1007 15:35:08.838673 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Oct 07 15:35:08 crc kubenswrapper[4959]: I1007 15:35:08.846735 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest"] Oct 07 15:35:09 crc kubenswrapper[4959]: I1007 15:35:09.024819 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g67fm\" (UniqueName: \"kubernetes.io/projected/f51247e0-f093-4d23-83a8-d8713b783834-kube-api-access-g67fm\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"f51247e0-f093-4d23-83a8-d8713b783834\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Oct 07 15:35:09 crc kubenswrapper[4959]: I1007 15:35:09.025026 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"f51247e0-f093-4d23-83a8-d8713b783834\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Oct 07 15:35:09 crc kubenswrapper[4959]: I1007 15:35:09.126497 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"f51247e0-f093-4d23-83a8-d8713b783834\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Oct 07 15:35:09 crc kubenswrapper[4959]: I1007 15:35:09.126660 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g67fm\" (UniqueName: \"kubernetes.io/projected/f51247e0-f093-4d23-83a8-d8713b783834-kube-api-access-g67fm\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"f51247e0-f093-4d23-83a8-d8713b783834\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Oct 07 15:35:09 crc kubenswrapper[4959]: I1007 15:35:09.127079 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"f51247e0-f093-4d23-83a8-d8713b783834\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Oct 07 15:35:09 crc kubenswrapper[4959]: I1007 15:35:09.152040 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g67fm\" (UniqueName: \"kubernetes.io/projected/f51247e0-f093-4d23-83a8-d8713b783834-kube-api-access-g67fm\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"f51247e0-f093-4d23-83a8-d8713b783834\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Oct 07 15:35:09 crc kubenswrapper[4959]: I1007 15:35:09.156872 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"f51247e0-f093-4d23-83a8-d8713b783834\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Oct 07 15:35:09 crc kubenswrapper[4959]: I1007 15:35:09.457716 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Oct 07 15:35:09 crc kubenswrapper[4959]: E1007 15:35:09.458300 4959 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 07 15:35:09 crc kubenswrapper[4959]: I1007 15:35:09.894054 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest"] Oct 07 15:35:09 crc kubenswrapper[4959]: W1007 15:35:09.903669 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf51247e0_f093_4d23_83a8_d8713b783834.slice/crio-79bb855a3e31cce760810001c28808ca8dbf8cf599b842fd4417b9dc68947528 WatchSource:0}: Error finding container 79bb855a3e31cce760810001c28808ca8dbf8cf599b842fd4417b9dc68947528: Status 404 returned error can't find the container with id 79bb855a3e31cce760810001c28808ca8dbf8cf599b842fd4417b9dc68947528 Oct 07 15:35:09 crc kubenswrapper[4959]: E1007 15:35:09.905583 4959 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 07 15:35:10 crc kubenswrapper[4959]: E1007 15:35:10.417497 4959 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 07 15:35:10 crc kubenswrapper[4959]: I1007 15:35:10.575876 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" event={"ID":"f51247e0-f093-4d23-83a8-d8713b783834","Type":"ContainerStarted","Data":"79bb855a3e31cce760810001c28808ca8dbf8cf599b842fd4417b9dc68947528"} Oct 07 15:35:11 crc kubenswrapper[4959]: I1007 15:35:11.587534 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" event={"ID":"f51247e0-f093-4d23-83a8-d8713b783834","Type":"ContainerStarted","Data":"65c696c47ef6c26d350af80c9d7b1a5a3c84e93626dc79473412680f816061ef"} Oct 07 15:35:11 crc kubenswrapper[4959]: E1007 15:35:11.588372 4959 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 07 15:35:11 crc kubenswrapper[4959]: I1007 15:35:11.602659 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" podStartSLOduration=3.092297043 podStartE2EDuration="3.602639704s" podCreationTimestamp="2025-10-07 15:35:08 +0000 UTC" firstStartedPulling="2025-10-07 15:35:09.907032007 +0000 UTC m=+6591.990436332" lastFinishedPulling="2025-10-07 15:35:10.417374678 +0000 UTC m=+6592.500778993" observedRunningTime="2025-10-07 15:35:11.600349546 +0000 UTC m=+6593.683753891" watchObservedRunningTime="2025-10-07 15:35:11.602639704 +0000 UTC m=+6593.686044029" Oct 07 15:35:12 crc kubenswrapper[4959]: E1007 15:35:12.598387 4959 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 07 15:35:19 crc kubenswrapper[4959]: I1007 15:35:19.653640 4959 scope.go:117] "RemoveContainer" containerID="02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a" Oct 07 15:35:19 crc kubenswrapper[4959]: E1007 15:35:19.654856 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:35:31 crc kubenswrapper[4959]: I1007 15:35:31.653568 4959 scope.go:117] "RemoveContainer" containerID="02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a" Oct 07 15:35:31 crc kubenswrapper[4959]: E1007 15:35:31.654484 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:35:42 crc kubenswrapper[4959]: I1007 15:35:42.653812 4959 scope.go:117] "RemoveContainer" containerID="02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a" Oct 07 15:35:42 crc kubenswrapper[4959]: E1007 15:35:42.654789 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:35:48 crc kubenswrapper[4959]: I1007 15:35:48.491124 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-tlcvp/must-gather-f4rxm"] Oct 07 15:35:48 crc kubenswrapper[4959]: I1007 15:35:48.493608 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tlcvp/must-gather-f4rxm" Oct 07 15:35:48 crc kubenswrapper[4959]: I1007 15:35:48.495405 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-tlcvp"/"default-dockercfg-jt7km" Oct 07 15:35:48 crc kubenswrapper[4959]: I1007 15:35:48.506502 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-tlcvp"/"openshift-service-ca.crt" Oct 07 15:35:48 crc kubenswrapper[4959]: I1007 15:35:48.522818 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-tlcvp"/"kube-root-ca.crt" Oct 07 15:35:48 crc kubenswrapper[4959]: I1007 15:35:48.548479 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-tlcvp/must-gather-f4rxm"] Oct 07 15:35:48 crc kubenswrapper[4959]: I1007 15:35:48.614567 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt7zb\" (UniqueName: \"kubernetes.io/projected/d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5-kube-api-access-qt7zb\") pod \"must-gather-f4rxm\" (UID: \"d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5\") " pod="openshift-must-gather-tlcvp/must-gather-f4rxm" Oct 07 15:35:48 crc kubenswrapper[4959]: I1007 15:35:48.614984 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5-must-gather-output\") pod \"must-gather-f4rxm\" (UID: \"d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5\") " pod="openshift-must-gather-tlcvp/must-gather-f4rxm" Oct 07 15:35:48 crc kubenswrapper[4959]: I1007 15:35:48.717387 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qt7zb\" (UniqueName: \"kubernetes.io/projected/d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5-kube-api-access-qt7zb\") pod \"must-gather-f4rxm\" (UID: \"d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5\") " pod="openshift-must-gather-tlcvp/must-gather-f4rxm" Oct 07 15:35:48 crc kubenswrapper[4959]: I1007 15:35:48.717838 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5-must-gather-output\") pod \"must-gather-f4rxm\" (UID: \"d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5\") " pod="openshift-must-gather-tlcvp/must-gather-f4rxm" Oct 07 15:35:48 crc kubenswrapper[4959]: I1007 15:35:48.718358 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5-must-gather-output\") pod \"must-gather-f4rxm\" (UID: \"d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5\") " pod="openshift-must-gather-tlcvp/must-gather-f4rxm" Oct 07 15:35:48 crc kubenswrapper[4959]: I1007 15:35:48.745315 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qt7zb\" (UniqueName: \"kubernetes.io/projected/d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5-kube-api-access-qt7zb\") pod \"must-gather-f4rxm\" (UID: \"d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5\") " pod="openshift-must-gather-tlcvp/must-gather-f4rxm" Oct 07 15:35:48 crc kubenswrapper[4959]: I1007 15:35:48.830592 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tlcvp/must-gather-f4rxm" Oct 07 15:35:49 crc kubenswrapper[4959]: I1007 15:35:49.470037 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-tlcvp/must-gather-f4rxm"] Oct 07 15:35:49 crc kubenswrapper[4959]: I1007 15:35:49.946784 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tlcvp/must-gather-f4rxm" event={"ID":"d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5","Type":"ContainerStarted","Data":"72999b4164f6c9dd6a1c4484932d099604ea98fb292f83085f77321b513d1769"} Oct 07 15:35:53 crc kubenswrapper[4959]: I1007 15:35:53.992712 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tlcvp/must-gather-f4rxm" event={"ID":"d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5","Type":"ContainerStarted","Data":"244c8cbefa7361365f17179daec007e4317aa8ef6c39427a8ba5e1719b8b0560"} Oct 07 15:35:54 crc kubenswrapper[4959]: I1007 15:35:54.654040 4959 scope.go:117] "RemoveContainer" containerID="02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a" Oct 07 15:35:54 crc kubenswrapper[4959]: E1007 15:35:54.654333 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:35:55 crc kubenswrapper[4959]: I1007 15:35:55.002817 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tlcvp/must-gather-f4rxm" event={"ID":"d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5","Type":"ContainerStarted","Data":"78053427f49238a92dfe17edbca923b5516edee46ae77d2aa0e841622bc9783b"} Oct 07 15:35:55 crc kubenswrapper[4959]: I1007 15:35:55.017658 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-tlcvp/must-gather-f4rxm" podStartSLOduration=2.936634033 podStartE2EDuration="7.017639574s" podCreationTimestamp="2025-10-07 15:35:48 +0000 UTC" firstStartedPulling="2025-10-07 15:35:49.483979319 +0000 UTC m=+6631.567383644" lastFinishedPulling="2025-10-07 15:35:53.56498486 +0000 UTC m=+6635.648389185" observedRunningTime="2025-10-07 15:35:55.014841313 +0000 UTC m=+6637.098245638" watchObservedRunningTime="2025-10-07 15:35:55.017639574 +0000 UTC m=+6637.101043899" Oct 07 15:35:58 crc kubenswrapper[4959]: I1007 15:35:58.986978 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-tlcvp/crc-debug-2c9dg"] Oct 07 15:35:58 crc kubenswrapper[4959]: I1007 15:35:58.989049 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tlcvp/crc-debug-2c9dg" Oct 07 15:35:59 crc kubenswrapper[4959]: I1007 15:35:59.163513 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zth8w\" (UniqueName: \"kubernetes.io/projected/7a76f0ab-a8cf-4470-bc5f-09e711ba3fee-kube-api-access-zth8w\") pod \"crc-debug-2c9dg\" (UID: \"7a76f0ab-a8cf-4470-bc5f-09e711ba3fee\") " pod="openshift-must-gather-tlcvp/crc-debug-2c9dg" Oct 07 15:35:59 crc kubenswrapper[4959]: I1007 15:35:59.163608 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7a76f0ab-a8cf-4470-bc5f-09e711ba3fee-host\") pod \"crc-debug-2c9dg\" (UID: \"7a76f0ab-a8cf-4470-bc5f-09e711ba3fee\") " pod="openshift-must-gather-tlcvp/crc-debug-2c9dg" Oct 07 15:35:59 crc kubenswrapper[4959]: I1007 15:35:59.265871 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zth8w\" (UniqueName: \"kubernetes.io/projected/7a76f0ab-a8cf-4470-bc5f-09e711ba3fee-kube-api-access-zth8w\") pod \"crc-debug-2c9dg\" (UID: \"7a76f0ab-a8cf-4470-bc5f-09e711ba3fee\") " pod="openshift-must-gather-tlcvp/crc-debug-2c9dg" Oct 07 15:35:59 crc kubenswrapper[4959]: I1007 15:35:59.265998 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7a76f0ab-a8cf-4470-bc5f-09e711ba3fee-host\") pod \"crc-debug-2c9dg\" (UID: \"7a76f0ab-a8cf-4470-bc5f-09e711ba3fee\") " pod="openshift-must-gather-tlcvp/crc-debug-2c9dg" Oct 07 15:35:59 crc kubenswrapper[4959]: I1007 15:35:59.266218 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7a76f0ab-a8cf-4470-bc5f-09e711ba3fee-host\") pod \"crc-debug-2c9dg\" (UID: \"7a76f0ab-a8cf-4470-bc5f-09e711ba3fee\") " pod="openshift-must-gather-tlcvp/crc-debug-2c9dg" Oct 07 15:35:59 crc kubenswrapper[4959]: I1007 15:35:59.293562 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zth8w\" (UniqueName: \"kubernetes.io/projected/7a76f0ab-a8cf-4470-bc5f-09e711ba3fee-kube-api-access-zth8w\") pod \"crc-debug-2c9dg\" (UID: \"7a76f0ab-a8cf-4470-bc5f-09e711ba3fee\") " pod="openshift-must-gather-tlcvp/crc-debug-2c9dg" Oct 07 15:35:59 crc kubenswrapper[4959]: I1007 15:35:59.313611 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tlcvp/crc-debug-2c9dg" Oct 07 15:35:59 crc kubenswrapper[4959]: W1007 15:35:59.358444 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a76f0ab_a8cf_4470_bc5f_09e711ba3fee.slice/crio-38544a16af79ae286916b8184e3a009ae2ed330206f51dedb10c473fe54ac448 WatchSource:0}: Error finding container 38544a16af79ae286916b8184e3a009ae2ed330206f51dedb10c473fe54ac448: Status 404 returned error can't find the container with id 38544a16af79ae286916b8184e3a009ae2ed330206f51dedb10c473fe54ac448 Oct 07 15:36:00 crc kubenswrapper[4959]: I1007 15:36:00.053039 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tlcvp/crc-debug-2c9dg" event={"ID":"7a76f0ab-a8cf-4470-bc5f-09e711ba3fee","Type":"ContainerStarted","Data":"38544a16af79ae286916b8184e3a009ae2ed330206f51dedb10c473fe54ac448"} Oct 07 15:36:07 crc kubenswrapper[4959]: I1007 15:36:07.653987 4959 scope.go:117] "RemoveContainer" containerID="02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a" Oct 07 15:36:07 crc kubenswrapper[4959]: E1007 15:36:07.654770 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:36:13 crc kubenswrapper[4959]: I1007 15:36:13.192891 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tlcvp/crc-debug-2c9dg" event={"ID":"7a76f0ab-a8cf-4470-bc5f-09e711ba3fee","Type":"ContainerStarted","Data":"e2239caf5f94b699aa8d89234db94aa42555f219249ef40a61f0d89eab4dcf1b"} Oct 07 15:36:13 crc kubenswrapper[4959]: I1007 15:36:13.214237 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-tlcvp/crc-debug-2c9dg" podStartSLOduration=1.806940783 podStartE2EDuration="15.214210501s" podCreationTimestamp="2025-10-07 15:35:58 +0000 UTC" firstStartedPulling="2025-10-07 15:35:59.361169219 +0000 UTC m=+6641.444573564" lastFinishedPulling="2025-10-07 15:36:12.768438957 +0000 UTC m=+6654.851843282" observedRunningTime="2025-10-07 15:36:13.210839365 +0000 UTC m=+6655.294243690" watchObservedRunningTime="2025-10-07 15:36:13.214210501 +0000 UTC m=+6655.297614836" Oct 07 15:36:19 crc kubenswrapper[4959]: I1007 15:36:19.653119 4959 scope.go:117] "RemoveContainer" containerID="02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a" Oct 07 15:36:19 crc kubenswrapper[4959]: E1007 15:36:19.654035 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:36:32 crc kubenswrapper[4959]: I1007 15:36:32.653858 4959 scope.go:117] "RemoveContainer" containerID="02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a" Oct 07 15:36:32 crc kubenswrapper[4959]: E1007 15:36:32.654770 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:36:34 crc kubenswrapper[4959]: E1007 15:36:34.653499 4959 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 07 15:36:44 crc kubenswrapper[4959]: I1007 15:36:44.654756 4959 scope.go:117] "RemoveContainer" containerID="02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a" Oct 07 15:36:44 crc kubenswrapper[4959]: E1007 15:36:44.656078 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:36:49 crc kubenswrapper[4959]: I1007 15:36:49.224218 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xlbrz"] Oct 07 15:36:49 crc kubenswrapper[4959]: I1007 15:36:49.228179 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xlbrz" Oct 07 15:36:49 crc kubenswrapper[4959]: I1007 15:36:49.249957 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xlbrz"] Oct 07 15:36:49 crc kubenswrapper[4959]: I1007 15:36:49.250184 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce3552bd-baa1-4b76-9943-8cc375c11945-catalog-content\") pod \"certified-operators-xlbrz\" (UID: \"ce3552bd-baa1-4b76-9943-8cc375c11945\") " pod="openshift-marketplace/certified-operators-xlbrz" Oct 07 15:36:49 crc kubenswrapper[4959]: I1007 15:36:49.250267 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce3552bd-baa1-4b76-9943-8cc375c11945-utilities\") pod \"certified-operators-xlbrz\" (UID: \"ce3552bd-baa1-4b76-9943-8cc375c11945\") " pod="openshift-marketplace/certified-operators-xlbrz" Oct 07 15:36:49 crc kubenswrapper[4959]: I1007 15:36:49.250313 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqwfh\" (UniqueName: \"kubernetes.io/projected/ce3552bd-baa1-4b76-9943-8cc375c11945-kube-api-access-cqwfh\") pod \"certified-operators-xlbrz\" (UID: \"ce3552bd-baa1-4b76-9943-8cc375c11945\") " pod="openshift-marketplace/certified-operators-xlbrz" Oct 07 15:36:49 crc kubenswrapper[4959]: I1007 15:36:49.354870 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce3552bd-baa1-4b76-9943-8cc375c11945-catalog-content\") pod \"certified-operators-xlbrz\" (UID: \"ce3552bd-baa1-4b76-9943-8cc375c11945\") " pod="openshift-marketplace/certified-operators-xlbrz" Oct 07 15:36:49 crc kubenswrapper[4959]: I1007 15:36:49.354942 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce3552bd-baa1-4b76-9943-8cc375c11945-utilities\") pod \"certified-operators-xlbrz\" (UID: \"ce3552bd-baa1-4b76-9943-8cc375c11945\") " pod="openshift-marketplace/certified-operators-xlbrz" Oct 07 15:36:49 crc kubenswrapper[4959]: I1007 15:36:49.354986 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqwfh\" (UniqueName: \"kubernetes.io/projected/ce3552bd-baa1-4b76-9943-8cc375c11945-kube-api-access-cqwfh\") pod \"certified-operators-xlbrz\" (UID: \"ce3552bd-baa1-4b76-9943-8cc375c11945\") " pod="openshift-marketplace/certified-operators-xlbrz" Oct 07 15:36:49 crc kubenswrapper[4959]: I1007 15:36:49.356030 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce3552bd-baa1-4b76-9943-8cc375c11945-catalog-content\") pod \"certified-operators-xlbrz\" (UID: \"ce3552bd-baa1-4b76-9943-8cc375c11945\") " pod="openshift-marketplace/certified-operators-xlbrz" Oct 07 15:36:49 crc kubenswrapper[4959]: I1007 15:36:49.356299 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce3552bd-baa1-4b76-9943-8cc375c11945-utilities\") pod \"certified-operators-xlbrz\" (UID: \"ce3552bd-baa1-4b76-9943-8cc375c11945\") " pod="openshift-marketplace/certified-operators-xlbrz" Oct 07 15:36:49 crc kubenswrapper[4959]: I1007 15:36:49.377909 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqwfh\" (UniqueName: \"kubernetes.io/projected/ce3552bd-baa1-4b76-9943-8cc375c11945-kube-api-access-cqwfh\") pod \"certified-operators-xlbrz\" (UID: \"ce3552bd-baa1-4b76-9943-8cc375c11945\") " pod="openshift-marketplace/certified-operators-xlbrz" Oct 07 15:36:49 crc kubenswrapper[4959]: I1007 15:36:49.565755 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xlbrz" Oct 07 15:36:50 crc kubenswrapper[4959]: I1007 15:36:50.434834 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xlbrz"] Oct 07 15:36:50 crc kubenswrapper[4959]: I1007 15:36:50.603692 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xlbrz" event={"ID":"ce3552bd-baa1-4b76-9943-8cc375c11945","Type":"ContainerStarted","Data":"5402cb8722749d32914a0668ad9d96dffee76272564f18fcdc80082e59a472ba"} Oct 07 15:36:51 crc kubenswrapper[4959]: I1007 15:36:51.614972 4959 generic.go:334] "Generic (PLEG): container finished" podID="ce3552bd-baa1-4b76-9943-8cc375c11945" containerID="2ce46db11147b531826c4ec28cb1d4bf00a60cc1e6c8a32f3bfd6692d1b93789" exitCode=0 Oct 07 15:36:51 crc kubenswrapper[4959]: I1007 15:36:51.615173 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xlbrz" event={"ID":"ce3552bd-baa1-4b76-9943-8cc375c11945","Type":"ContainerDied","Data":"2ce46db11147b531826c4ec28cb1d4bf00a60cc1e6c8a32f3bfd6692d1b93789"} Oct 07 15:36:51 crc kubenswrapper[4959]: I1007 15:36:51.622696 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 07 15:36:53 crc kubenswrapper[4959]: I1007 15:36:53.638054 4959 generic.go:334] "Generic (PLEG): container finished" podID="ce3552bd-baa1-4b76-9943-8cc375c11945" containerID="78c2bea59f3f9eac86300c98d87189bd67e4acf542ad2b3f366463914b2cc7e4" exitCode=0 Oct 07 15:36:53 crc kubenswrapper[4959]: I1007 15:36:53.638168 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xlbrz" event={"ID":"ce3552bd-baa1-4b76-9943-8cc375c11945","Type":"ContainerDied","Data":"78c2bea59f3f9eac86300c98d87189bd67e4acf542ad2b3f366463914b2cc7e4"} Oct 07 15:36:54 crc kubenswrapper[4959]: I1007 15:36:54.219378 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zftzn"] Oct 07 15:36:54 crc kubenswrapper[4959]: I1007 15:36:54.223710 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zftzn" Oct 07 15:36:54 crc kubenswrapper[4959]: I1007 15:36:54.231374 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zftzn"] Oct 07 15:36:54 crc kubenswrapper[4959]: I1007 15:36:54.390132 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mpzh\" (UniqueName: \"kubernetes.io/projected/e56a45bf-949b-4c96-9978-c8929478ebb7-kube-api-access-7mpzh\") pod \"redhat-operators-zftzn\" (UID: \"e56a45bf-949b-4c96-9978-c8929478ebb7\") " pod="openshift-marketplace/redhat-operators-zftzn" Oct 07 15:36:54 crc kubenswrapper[4959]: I1007 15:36:54.390683 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e56a45bf-949b-4c96-9978-c8929478ebb7-utilities\") pod \"redhat-operators-zftzn\" (UID: \"e56a45bf-949b-4c96-9978-c8929478ebb7\") " pod="openshift-marketplace/redhat-operators-zftzn" Oct 07 15:36:54 crc kubenswrapper[4959]: I1007 15:36:54.390869 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e56a45bf-949b-4c96-9978-c8929478ebb7-catalog-content\") pod \"redhat-operators-zftzn\" (UID: \"e56a45bf-949b-4c96-9978-c8929478ebb7\") " pod="openshift-marketplace/redhat-operators-zftzn" Oct 07 15:36:54 crc kubenswrapper[4959]: I1007 15:36:54.493032 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e56a45bf-949b-4c96-9978-c8929478ebb7-catalog-content\") pod \"redhat-operators-zftzn\" (UID: \"e56a45bf-949b-4c96-9978-c8929478ebb7\") " pod="openshift-marketplace/redhat-operators-zftzn" Oct 07 15:36:54 crc kubenswrapper[4959]: I1007 15:36:54.493181 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mpzh\" (UniqueName: \"kubernetes.io/projected/e56a45bf-949b-4c96-9978-c8929478ebb7-kube-api-access-7mpzh\") pod \"redhat-operators-zftzn\" (UID: \"e56a45bf-949b-4c96-9978-c8929478ebb7\") " pod="openshift-marketplace/redhat-operators-zftzn" Oct 07 15:36:54 crc kubenswrapper[4959]: I1007 15:36:54.493206 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e56a45bf-949b-4c96-9978-c8929478ebb7-utilities\") pod \"redhat-operators-zftzn\" (UID: \"e56a45bf-949b-4c96-9978-c8929478ebb7\") " pod="openshift-marketplace/redhat-operators-zftzn" Oct 07 15:36:54 crc kubenswrapper[4959]: I1007 15:36:54.493782 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e56a45bf-949b-4c96-9978-c8929478ebb7-catalog-content\") pod \"redhat-operators-zftzn\" (UID: \"e56a45bf-949b-4c96-9978-c8929478ebb7\") " pod="openshift-marketplace/redhat-operators-zftzn" Oct 07 15:36:54 crc kubenswrapper[4959]: I1007 15:36:54.493794 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e56a45bf-949b-4c96-9978-c8929478ebb7-utilities\") pod \"redhat-operators-zftzn\" (UID: \"e56a45bf-949b-4c96-9978-c8929478ebb7\") " pod="openshift-marketplace/redhat-operators-zftzn" Oct 07 15:36:54 crc kubenswrapper[4959]: I1007 15:36:54.520315 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mpzh\" (UniqueName: \"kubernetes.io/projected/e56a45bf-949b-4c96-9978-c8929478ebb7-kube-api-access-7mpzh\") pod \"redhat-operators-zftzn\" (UID: \"e56a45bf-949b-4c96-9978-c8929478ebb7\") " pod="openshift-marketplace/redhat-operators-zftzn" Oct 07 15:36:54 crc kubenswrapper[4959]: I1007 15:36:54.574631 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zftzn" Oct 07 15:36:54 crc kubenswrapper[4959]: I1007 15:36:54.684060 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xlbrz" event={"ID":"ce3552bd-baa1-4b76-9943-8cc375c11945","Type":"ContainerStarted","Data":"f03f154e86dcf3b8befff10555641e6615ee1bfec79eb2eb30c5a804fe645201"} Oct 07 15:36:54 crc kubenswrapper[4959]: I1007 15:36:54.704791 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xlbrz" podStartSLOduration=3.202338031 podStartE2EDuration="5.704770489s" podCreationTimestamp="2025-10-07 15:36:49 +0000 UTC" firstStartedPulling="2025-10-07 15:36:51.622366796 +0000 UTC m=+6693.705771121" lastFinishedPulling="2025-10-07 15:36:54.124799254 +0000 UTC m=+6696.208203579" observedRunningTime="2025-10-07 15:36:54.698070738 +0000 UTC m=+6696.781475063" watchObservedRunningTime="2025-10-07 15:36:54.704770489 +0000 UTC m=+6696.788174814" Oct 07 15:36:55 crc kubenswrapper[4959]: I1007 15:36:55.107781 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zftzn"] Oct 07 15:36:55 crc kubenswrapper[4959]: W1007 15:36:55.112942 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode56a45bf_949b_4c96_9978_c8929478ebb7.slice/crio-43f8842967a6e5870d6a58052a90b371e1038192482bb74e277df6b4d708115d WatchSource:0}: Error finding container 43f8842967a6e5870d6a58052a90b371e1038192482bb74e277df6b4d708115d: Status 404 returned error can't find the container with id 43f8842967a6e5870d6a58052a90b371e1038192482bb74e277df6b4d708115d Oct 07 15:36:55 crc kubenswrapper[4959]: I1007 15:36:55.695906 4959 generic.go:334] "Generic (PLEG): container finished" podID="e56a45bf-949b-4c96-9978-c8929478ebb7" containerID="116bc2a4db4b8579eceff3324967b6567bebd4250dcb44672be872e2497a514a" exitCode=0 Oct 07 15:36:55 crc kubenswrapper[4959]: I1007 15:36:55.696020 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zftzn" event={"ID":"e56a45bf-949b-4c96-9978-c8929478ebb7","Type":"ContainerDied","Data":"116bc2a4db4b8579eceff3324967b6567bebd4250dcb44672be872e2497a514a"} Oct 07 15:36:55 crc kubenswrapper[4959]: I1007 15:36:55.696989 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zftzn" event={"ID":"e56a45bf-949b-4c96-9978-c8929478ebb7","Type":"ContainerStarted","Data":"43f8842967a6e5870d6a58052a90b371e1038192482bb74e277df6b4d708115d"} Oct 07 15:36:57 crc kubenswrapper[4959]: I1007 15:36:57.654567 4959 scope.go:117] "RemoveContainer" containerID="02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a" Oct 07 15:36:57 crc kubenswrapper[4959]: E1007 15:36:57.655759 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:36:59 crc kubenswrapper[4959]: I1007 15:36:59.566440 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xlbrz" Oct 07 15:37:00 crc kubenswrapper[4959]: I1007 15:36:59.568394 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xlbrz" Oct 07 15:37:00 crc kubenswrapper[4959]: I1007 15:36:59.639717 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xlbrz" Oct 07 15:37:00 crc kubenswrapper[4959]: I1007 15:36:59.822923 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xlbrz" Oct 07 15:37:05 crc kubenswrapper[4959]: I1007 15:37:05.209949 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xlbrz"] Oct 07 15:37:05 crc kubenswrapper[4959]: I1007 15:37:05.210856 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xlbrz" podUID="ce3552bd-baa1-4b76-9943-8cc375c11945" containerName="registry-server" containerID="cri-o://f03f154e86dcf3b8befff10555641e6615ee1bfec79eb2eb30c5a804fe645201" gracePeriod=2 Oct 07 15:37:05 crc kubenswrapper[4959]: I1007 15:37:05.820635 4959 generic.go:334] "Generic (PLEG): container finished" podID="ce3552bd-baa1-4b76-9943-8cc375c11945" containerID="f03f154e86dcf3b8befff10555641e6615ee1bfec79eb2eb30c5a804fe645201" exitCode=0 Oct 07 15:37:05 crc kubenswrapper[4959]: I1007 15:37:05.820690 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xlbrz" event={"ID":"ce3552bd-baa1-4b76-9943-8cc375c11945","Type":"ContainerDied","Data":"f03f154e86dcf3b8befff10555641e6615ee1bfec79eb2eb30c5a804fe645201"} Oct 07 15:37:08 crc kubenswrapper[4959]: I1007 15:37:08.722468 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xlbrz" Oct 07 15:37:08 crc kubenswrapper[4959]: I1007 15:37:08.850010 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xlbrz" event={"ID":"ce3552bd-baa1-4b76-9943-8cc375c11945","Type":"ContainerDied","Data":"5402cb8722749d32914a0668ad9d96dffee76272564f18fcdc80082e59a472ba"} Oct 07 15:37:08 crc kubenswrapper[4959]: I1007 15:37:08.850351 4959 scope.go:117] "RemoveContainer" containerID="f03f154e86dcf3b8befff10555641e6615ee1bfec79eb2eb30c5a804fe645201" Oct 07 15:37:08 crc kubenswrapper[4959]: I1007 15:37:08.850317 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xlbrz" Oct 07 15:37:08 crc kubenswrapper[4959]: I1007 15:37:08.853052 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce3552bd-baa1-4b76-9943-8cc375c11945-utilities\") pod \"ce3552bd-baa1-4b76-9943-8cc375c11945\" (UID: \"ce3552bd-baa1-4b76-9943-8cc375c11945\") " Oct 07 15:37:08 crc kubenswrapper[4959]: I1007 15:37:08.853138 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqwfh\" (UniqueName: \"kubernetes.io/projected/ce3552bd-baa1-4b76-9943-8cc375c11945-kube-api-access-cqwfh\") pod \"ce3552bd-baa1-4b76-9943-8cc375c11945\" (UID: \"ce3552bd-baa1-4b76-9943-8cc375c11945\") " Oct 07 15:37:08 crc kubenswrapper[4959]: I1007 15:37:08.853163 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce3552bd-baa1-4b76-9943-8cc375c11945-catalog-content\") pod \"ce3552bd-baa1-4b76-9943-8cc375c11945\" (UID: \"ce3552bd-baa1-4b76-9943-8cc375c11945\") " Oct 07 15:37:08 crc kubenswrapper[4959]: I1007 15:37:08.853482 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zftzn" event={"ID":"e56a45bf-949b-4c96-9978-c8929478ebb7","Type":"ContainerStarted","Data":"9ba78b351c810863bab74522b813373ef5b8539bda33a109ebc750d7142be82b"} Oct 07 15:37:08 crc kubenswrapper[4959]: I1007 15:37:08.854462 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce3552bd-baa1-4b76-9943-8cc375c11945-utilities" (OuterVolumeSpecName: "utilities") pod "ce3552bd-baa1-4b76-9943-8cc375c11945" (UID: "ce3552bd-baa1-4b76-9943-8cc375c11945"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:37:08 crc kubenswrapper[4959]: I1007 15:37:08.885611 4959 scope.go:117] "RemoveContainer" containerID="78c2bea59f3f9eac86300c98d87189bd67e4acf542ad2b3f366463914b2cc7e4" Oct 07 15:37:08 crc kubenswrapper[4959]: I1007 15:37:08.885649 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce3552bd-baa1-4b76-9943-8cc375c11945-kube-api-access-cqwfh" (OuterVolumeSpecName: "kube-api-access-cqwfh") pod "ce3552bd-baa1-4b76-9943-8cc375c11945" (UID: "ce3552bd-baa1-4b76-9943-8cc375c11945"). InnerVolumeSpecName "kube-api-access-cqwfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 15:37:08 crc kubenswrapper[4959]: I1007 15:37:08.952761 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce3552bd-baa1-4b76-9943-8cc375c11945-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ce3552bd-baa1-4b76-9943-8cc375c11945" (UID: "ce3552bd-baa1-4b76-9943-8cc375c11945"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:37:08 crc kubenswrapper[4959]: I1007 15:37:08.955632 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce3552bd-baa1-4b76-9943-8cc375c11945-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 15:37:08 crc kubenswrapper[4959]: I1007 15:37:08.955674 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqwfh\" (UniqueName: \"kubernetes.io/projected/ce3552bd-baa1-4b76-9943-8cc375c11945-kube-api-access-cqwfh\") on node \"crc\" DevicePath \"\"" Oct 07 15:37:08 crc kubenswrapper[4959]: I1007 15:37:08.955688 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce3552bd-baa1-4b76-9943-8cc375c11945-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 15:37:08 crc kubenswrapper[4959]: I1007 15:37:08.973971 4959 scope.go:117] "RemoveContainer" containerID="2ce46db11147b531826c4ec28cb1d4bf00a60cc1e6c8a32f3bfd6692d1b93789" Oct 07 15:37:09 crc kubenswrapper[4959]: I1007 15:37:09.194498 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xlbrz"] Oct 07 15:37:09 crc kubenswrapper[4959]: I1007 15:37:09.205282 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xlbrz"] Oct 07 15:37:09 crc kubenswrapper[4959]: I1007 15:37:09.868865 4959 generic.go:334] "Generic (PLEG): container finished" podID="e56a45bf-949b-4c96-9978-c8929478ebb7" containerID="9ba78b351c810863bab74522b813373ef5b8539bda33a109ebc750d7142be82b" exitCode=0 Oct 07 15:37:09 crc kubenswrapper[4959]: I1007 15:37:09.869374 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zftzn" event={"ID":"e56a45bf-949b-4c96-9978-c8929478ebb7","Type":"ContainerDied","Data":"9ba78b351c810863bab74522b813373ef5b8539bda33a109ebc750d7142be82b"} Oct 07 15:37:10 crc kubenswrapper[4959]: I1007 15:37:10.653784 4959 scope.go:117] "RemoveContainer" containerID="02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a" Oct 07 15:37:10 crc kubenswrapper[4959]: E1007 15:37:10.654630 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:37:10 crc kubenswrapper[4959]: I1007 15:37:10.670434 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce3552bd-baa1-4b76-9943-8cc375c11945" path="/var/lib/kubelet/pods/ce3552bd-baa1-4b76-9943-8cc375c11945/volumes" Oct 07 15:37:11 crc kubenswrapper[4959]: I1007 15:37:11.733563 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ansibletest-ansibletest_954ed5c7-ad7c-4cd0-ac69-ffa0c9fa3c34/ansibletest-ansibletest/0.log" Oct 07 15:37:11 crc kubenswrapper[4959]: I1007 15:37:11.914511 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zftzn" event={"ID":"e56a45bf-949b-4c96-9978-c8929478ebb7","Type":"ContainerStarted","Data":"b00ee24996d4b3815b8d8127a8e35b7daf011db27be3d2896a0de44dc6b9e0fb"} Oct 07 15:37:11 crc kubenswrapper[4959]: I1007 15:37:11.946072 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zftzn" podStartSLOduration=2.894909975 podStartE2EDuration="17.946039162s" podCreationTimestamp="2025-10-07 15:36:54 +0000 UTC" firstStartedPulling="2025-10-07 15:36:55.704273041 +0000 UTC m=+6697.787677366" lastFinishedPulling="2025-10-07 15:37:10.755402228 +0000 UTC m=+6712.838806553" observedRunningTime="2025-10-07 15:37:11.937487914 +0000 UTC m=+6714.020892249" watchObservedRunningTime="2025-10-07 15:37:11.946039162 +0000 UTC m=+6714.029443487" Oct 07 15:37:11 crc kubenswrapper[4959]: I1007 15:37:11.983598 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-646c6bbb48-kcmnj_3a0d41b9-b996-4597-8646-4fb61c6dc354/barbican-api/0.log" Oct 07 15:37:12 crc kubenswrapper[4959]: I1007 15:37:12.210631 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-646c6bbb48-kcmnj_3a0d41b9-b996-4597-8646-4fb61c6dc354/barbican-api-log/0.log" Oct 07 15:37:12 crc kubenswrapper[4959]: I1007 15:37:12.473954 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-76c697766b-9qfh8_a4a56ff7-04cc-48e9-be4a-651a98c06204/barbican-keystone-listener/0.log" Oct 07 15:37:13 crc kubenswrapper[4959]: I1007 15:37:13.052857 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7748b8ffbf-c9js9_d4f04734-09dc-47cf-9395-5f2ca7739a3f/barbican-worker/0.log" Oct 07 15:37:13 crc kubenswrapper[4959]: I1007 15:37:13.060201 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-76c697766b-9qfh8_a4a56ff7-04cc-48e9-be4a-651a98c06204/barbican-keystone-listener-log/0.log" Oct 07 15:37:13 crc kubenswrapper[4959]: I1007 15:37:13.335024 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7748b8ffbf-c9js9_d4f04734-09dc-47cf-9395-5f2ca7739a3f/barbican-worker-log/0.log" Oct 07 15:37:13 crc kubenswrapper[4959]: I1007 15:37:13.600745 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-gwdpl_4ddf979d-a363-420d-8368-aac337f5b078/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 15:37:13 crc kubenswrapper[4959]: I1007 15:37:13.871994 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0cc2047c-77e3-47f7-ae32-0053a6a17d25/ceilometer-central-agent/0.log" Oct 07 15:37:13 crc kubenswrapper[4959]: I1007 15:37:13.971937 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0cc2047c-77e3-47f7-ae32-0053a6a17d25/ceilometer-notification-agent/0.log" Oct 07 15:37:14 crc kubenswrapper[4959]: I1007 15:37:14.092290 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0cc2047c-77e3-47f7-ae32-0053a6a17d25/proxy-httpd/0.log" Oct 07 15:37:14 crc kubenswrapper[4959]: I1007 15:37:14.282422 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0cc2047c-77e3-47f7-ae32-0053a6a17d25/sg-core/0.log" Oct 07 15:37:14 crc kubenswrapper[4959]: I1007 15:37:14.554269 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-h6fzl_defafb0d-5ea5-40f6-9918-4c97a946219c/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 15:37:14 crc kubenswrapper[4959]: I1007 15:37:14.574893 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zftzn" Oct 07 15:37:14 crc kubenswrapper[4959]: I1007 15:37:14.574984 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zftzn" Oct 07 15:37:14 crc kubenswrapper[4959]: I1007 15:37:14.708283 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lb4bj_2325565a-d2ae-45a1-bece-46cda83edc36/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 15:37:14 crc kubenswrapper[4959]: I1007 15:37:14.993846 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_275d4f42-616d-432a-b4b3-932f4df2ab66/cinder-api/0.log" Oct 07 15:37:15 crc kubenswrapper[4959]: I1007 15:37:15.037260 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_275d4f42-616d-432a-b4b3-932f4df2ab66/cinder-api-log/0.log" Oct 07 15:37:15 crc kubenswrapper[4959]: I1007 15:37:15.330211 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_ef75c9c7-34ea-49a0-94f2-9182065a05c0/cinder-backup/0.log" Oct 07 15:37:15 crc kubenswrapper[4959]: I1007 15:37:15.355864 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_ef75c9c7-34ea-49a0-94f2-9182065a05c0/probe/0.log" Oct 07 15:37:15 crc kubenswrapper[4959]: I1007 15:37:15.460463 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_126adfe4-0ab1-4952-9a03-526fb74cad41/cinder-scheduler/0.log" Oct 07 15:37:15 crc kubenswrapper[4959]: I1007 15:37:15.635838 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zftzn" podUID="e56a45bf-949b-4c96-9978-c8929478ebb7" containerName="registry-server" probeResult="failure" output=< Oct 07 15:37:15 crc kubenswrapper[4959]: timeout: failed to connect service ":50051" within 1s Oct 07 15:37:15 crc kubenswrapper[4959]: > Oct 07 15:37:15 crc kubenswrapper[4959]: I1007 15:37:15.689470 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_126adfe4-0ab1-4952-9a03-526fb74cad41/probe/0.log" Oct 07 15:37:15 crc kubenswrapper[4959]: I1007 15:37:15.719771 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_04f1296f-0cd0-4189-9f0d-81da20535c37/cinder-volume/0.log" Oct 07 15:37:15 crc kubenswrapper[4959]: I1007 15:37:15.778632 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_04f1296f-0cd0-4189-9f0d-81da20535c37/probe/0.log" Oct 07 15:37:16 crc kubenswrapper[4959]: I1007 15:37:16.036507 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-khd49_7fa6bf86-611f-4c87-8475-ac0c57f5bf0a/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 15:37:16 crc kubenswrapper[4959]: I1007 15:37:16.097198 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-v6tfv_57e3b6ad-f589-41cc-8ba5-8009fe0ab434/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 15:37:16 crc kubenswrapper[4959]: I1007 15:37:16.491021 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79d4ff65f9-n8c9v_d46df3be-9794-4019-8099-9c3757b5b468/init/0.log" Oct 07 15:37:16 crc kubenswrapper[4959]: I1007 15:37:16.722984 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79d4ff65f9-n8c9v_d46df3be-9794-4019-8099-9c3757b5b468/init/0.log" Oct 07 15:37:16 crc kubenswrapper[4959]: I1007 15:37:16.774379 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79d4ff65f9-n8c9v_d46df3be-9794-4019-8099-9c3757b5b468/dnsmasq-dns/0.log" Oct 07 15:37:16 crc kubenswrapper[4959]: I1007 15:37:16.888255 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_43b9df1d-3d35-45f9-bbad-f39f9d33c1db/glance-httpd/0.log" Oct 07 15:37:16 crc kubenswrapper[4959]: I1007 15:37:16.943367 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_43b9df1d-3d35-45f9-bbad-f39f9d33c1db/glance-log/0.log" Oct 07 15:37:17 crc kubenswrapper[4959]: I1007 15:37:17.084472 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_14f63797-4c08-4844-a4e6-7075dffc801c/glance-httpd/0.log" Oct 07 15:37:17 crc kubenswrapper[4959]: I1007 15:37:17.149019 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_14f63797-4c08-4844-a4e6-7075dffc801c/glance-log/0.log" Oct 07 15:37:17 crc kubenswrapper[4959]: I1007 15:37:17.501655 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-76d6fc6cbd-2g7sq_13b27eed-f50f-4474-ace5-0f12e733f6cf/horizon/0.log" Oct 07 15:37:17 crc kubenswrapper[4959]: I1007 15:37:17.589143 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizontest-tests-horizontest_8ae1c80c-a359-4751-80e1-58da5918db21/horizontest-tests-horizontest/0.log" Oct 07 15:37:17 crc kubenswrapper[4959]: I1007 15:37:17.770800 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-t6m6v_d2f04a89-92db-4fae-994a-dc901fa21909/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 15:37:18 crc kubenswrapper[4959]: I1007 15:37:18.068127 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-bdlsj_865d8bbd-aece-4d0f-b948-a611ebe0ed6b/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 15:37:18 crc kubenswrapper[4959]: I1007 15:37:18.459387 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29330821-dpxgj_e167acc8-aaaa-4595-8543-a9622ee92fc8/keystone-cron/0.log" Oct 07 15:37:18 crc kubenswrapper[4959]: I1007 15:37:18.561092 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-76d6fc6cbd-2g7sq_13b27eed-f50f-4474-ace5-0f12e733f6cf/horizon-log/0.log" Oct 07 15:37:18 crc kubenswrapper[4959]: I1007 15:37:18.772823 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_8b8dd772-3ff0-474c-8571-ddf2bed03a82/kube-state-metrics/0.log" Oct 07 15:37:19 crc kubenswrapper[4959]: I1007 15:37:19.097579 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-fswj9_e5aa9a4d-4a36-481d-8853-c8a5b586d974/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 15:37:19 crc kubenswrapper[4959]: I1007 15:37:19.413790 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_23643eb4-c61b-4785-9a95-413794a0b756/manila-api/0.log" Oct 07 15:37:19 crc kubenswrapper[4959]: I1007 15:37:19.459419 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_23643eb4-c61b-4785-9a95-413794a0b756/manila-api-log/0.log" Oct 07 15:37:19 crc kubenswrapper[4959]: I1007 15:37:19.596420 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5c77f969b5-4w2xs_33e694c5-266c-40e1-a805-c174cd094645/keystone-api/0.log" Oct 07 15:37:19 crc kubenswrapper[4959]: I1007 15:37:19.681833 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_24f33673-13e4-437a-9d3e-6528a9b81c35/probe/0.log" Oct 07 15:37:19 crc kubenswrapper[4959]: I1007 15:37:19.801348 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_24f33673-13e4-437a-9d3e-6528a9b81c35/manila-scheduler/0.log" Oct 07 15:37:19 crc kubenswrapper[4959]: I1007 15:37:19.907092 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_59f4d3f2-ccf2-4724-aa24-2de193d2a2bc/manila-share/0.log" Oct 07 15:37:19 crc kubenswrapper[4959]: I1007 15:37:19.988843 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_59f4d3f2-ccf2-4724-aa24-2de193d2a2bc/probe/0.log" Oct 07 15:37:20 crc kubenswrapper[4959]: I1007 15:37:20.847761 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-k9dlb_6645e33c-3472-41cd-8f39-a702210bc860/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 15:37:20 crc kubenswrapper[4959]: I1007 15:37:20.893608 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-67d6886b4f-lz5gr_8f7cac26-af4a-4db4-b216-f71d41ca9c74/neutron-httpd/0.log" Oct 07 15:37:21 crc kubenswrapper[4959]: I1007 15:37:21.548446 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-67d6886b4f-lz5gr_8f7cac26-af4a-4db4-b216-f71d41ca9c74/neutron-api/0.log" Oct 07 15:37:22 crc kubenswrapper[4959]: I1007 15:37:22.779926 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_ab63a619-32d4-4999-9056-cde8181e86ab/nova-cell0-conductor-conductor/0.log" Oct 07 15:37:22 crc kubenswrapper[4959]: I1007 15:37:22.812281 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_557f7b70-2476-481e-afaf-fc346d47b007/nova-api-log/0.log" Oct 07 15:37:23 crc kubenswrapper[4959]: I1007 15:37:23.302511 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_850dd238-e869-4484-9d52-3c67dc241120/nova-cell1-conductor-conductor/0.log" Oct 07 15:37:23 crc kubenswrapper[4959]: I1007 15:37:23.817378 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_88e679d5-b7cb-4e49-9f01-ab4ceacd43f1/nova-cell1-novncproxy-novncproxy/0.log" Oct 07 15:37:23 crc kubenswrapper[4959]: I1007 15:37:23.836506 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_557f7b70-2476-481e-afaf-fc346d47b007/nova-api-api/0.log" Oct 07 15:37:24 crc kubenswrapper[4959]: I1007 15:37:24.153442 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-rzbsh_ce45a1f9-8a26-46d4-a413-c175c4d51ab4/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 15:37:24 crc kubenswrapper[4959]: I1007 15:37:24.331258 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_a046abc0-ec2e-42ac-ad49-e336824e9317/nova-metadata-log/0.log" Oct 07 15:37:24 crc kubenswrapper[4959]: I1007 15:37:24.635080 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zftzn" Oct 07 15:37:24 crc kubenswrapper[4959]: I1007 15:37:24.698178 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zftzn" Oct 07 15:37:25 crc kubenswrapper[4959]: I1007 15:37:25.196154 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_6af5bf69-0961-4d34-9cc6-3b04eb48bca1/mysql-bootstrap/0.log" Oct 07 15:37:25 crc kubenswrapper[4959]: I1007 15:37:25.238448 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_22fed6d1-09b4-4d80-8fde-d3759d994af1/nova-scheduler-scheduler/0.log" Oct 07 15:37:25 crc kubenswrapper[4959]: I1007 15:37:25.252446 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zftzn"] Oct 07 15:37:25 crc kubenswrapper[4959]: I1007 15:37:25.428486 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nwksc"] Oct 07 15:37:25 crc kubenswrapper[4959]: I1007 15:37:25.428733 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nwksc" podUID="fad6288b-9766-448d-9f91-374bc79257dd" containerName="registry-server" containerID="cri-o://6424b528e2636d9fd5db8a284452d6d3698a0710c9d3984ee3447099dbf624a2" gracePeriod=2 Oct 07 15:37:25 crc kubenswrapper[4959]: I1007 15:37:25.466781 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_6af5bf69-0961-4d34-9cc6-3b04eb48bca1/mysql-bootstrap/0.log" Oct 07 15:37:25 crc kubenswrapper[4959]: I1007 15:37:25.637452 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_6af5bf69-0961-4d34-9cc6-3b04eb48bca1/galera/0.log" Oct 07 15:37:25 crc kubenswrapper[4959]: I1007 15:37:25.654798 4959 scope.go:117] "RemoveContainer" containerID="02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a" Oct 07 15:37:25 crc kubenswrapper[4959]: E1007 15:37:25.655013 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:37:26 crc kubenswrapper[4959]: I1007 15:37:26.087995 4959 generic.go:334] "Generic (PLEG): container finished" podID="fad6288b-9766-448d-9f91-374bc79257dd" containerID="6424b528e2636d9fd5db8a284452d6d3698a0710c9d3984ee3447099dbf624a2" exitCode=0 Oct 07 15:37:26 crc kubenswrapper[4959]: I1007 15:37:26.089021 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nwksc" event={"ID":"fad6288b-9766-448d-9f91-374bc79257dd","Type":"ContainerDied","Data":"6424b528e2636d9fd5db8a284452d6d3698a0710c9d3984ee3447099dbf624a2"} Oct 07 15:37:26 crc kubenswrapper[4959]: I1007 15:37:26.089068 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nwksc" event={"ID":"fad6288b-9766-448d-9f91-374bc79257dd","Type":"ContainerDied","Data":"8b1d9d995a76d4186c146f31a9a45f5b674d9a6206c89a572c9ca26700969d5f"} Oct 07 15:37:26 crc kubenswrapper[4959]: I1007 15:37:26.089085 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b1d9d995a76d4186c146f31a9a45f5b674d9a6206c89a572c9ca26700969d5f" Oct 07 15:37:26 crc kubenswrapper[4959]: I1007 15:37:26.093972 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_fb9134cb-b4a5-42d6-b0ca-b4d3e1174625/mysql-bootstrap/0.log" Oct 07 15:37:26 crc kubenswrapper[4959]: I1007 15:37:26.133690 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nwksc" Oct 07 15:37:26 crc kubenswrapper[4959]: I1007 15:37:26.262744 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzbm7\" (UniqueName: \"kubernetes.io/projected/fad6288b-9766-448d-9f91-374bc79257dd-kube-api-access-jzbm7\") pod \"fad6288b-9766-448d-9f91-374bc79257dd\" (UID: \"fad6288b-9766-448d-9f91-374bc79257dd\") " Oct 07 15:37:26 crc kubenswrapper[4959]: I1007 15:37:26.262980 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fad6288b-9766-448d-9f91-374bc79257dd-utilities\") pod \"fad6288b-9766-448d-9f91-374bc79257dd\" (UID: \"fad6288b-9766-448d-9f91-374bc79257dd\") " Oct 07 15:37:26 crc kubenswrapper[4959]: I1007 15:37:26.263179 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fad6288b-9766-448d-9f91-374bc79257dd-catalog-content\") pod \"fad6288b-9766-448d-9f91-374bc79257dd\" (UID: \"fad6288b-9766-448d-9f91-374bc79257dd\") " Oct 07 15:37:26 crc kubenswrapper[4959]: I1007 15:37:26.265221 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fad6288b-9766-448d-9f91-374bc79257dd-utilities" (OuterVolumeSpecName: "utilities") pod "fad6288b-9766-448d-9f91-374bc79257dd" (UID: "fad6288b-9766-448d-9f91-374bc79257dd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:37:26 crc kubenswrapper[4959]: I1007 15:37:26.301210 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fad6288b-9766-448d-9f91-374bc79257dd-kube-api-access-jzbm7" (OuterVolumeSpecName: "kube-api-access-jzbm7") pod "fad6288b-9766-448d-9f91-374bc79257dd" (UID: "fad6288b-9766-448d-9f91-374bc79257dd"). InnerVolumeSpecName "kube-api-access-jzbm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 15:37:26 crc kubenswrapper[4959]: I1007 15:37:26.358048 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_fb9134cb-b4a5-42d6-b0ca-b4d3e1174625/mysql-bootstrap/0.log" Oct 07 15:37:26 crc kubenswrapper[4959]: I1007 15:37:26.365566 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzbm7\" (UniqueName: \"kubernetes.io/projected/fad6288b-9766-448d-9f91-374bc79257dd-kube-api-access-jzbm7\") on node \"crc\" DevicePath \"\"" Oct 07 15:37:26 crc kubenswrapper[4959]: I1007 15:37:26.365601 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fad6288b-9766-448d-9f91-374bc79257dd-utilities\") on node \"crc\" DevicePath \"\"" Oct 07 15:37:26 crc kubenswrapper[4959]: I1007 15:37:26.371587 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fad6288b-9766-448d-9f91-374bc79257dd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fad6288b-9766-448d-9f91-374bc79257dd" (UID: "fad6288b-9766-448d-9f91-374bc79257dd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:37:26 crc kubenswrapper[4959]: I1007 15:37:26.467011 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fad6288b-9766-448d-9f91-374bc79257dd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 07 15:37:26 crc kubenswrapper[4959]: I1007 15:37:26.495324 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_fb9134cb-b4a5-42d6-b0ca-b4d3e1174625/galera/0.log" Oct 07 15:37:26 crc kubenswrapper[4959]: I1007 15:37:26.918117 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_9ba3dd4a-23e6-4617-ae86-d0cc64bec9d8/openstackclient/0.log" Oct 07 15:37:27 crc kubenswrapper[4959]: I1007 15:37:27.098851 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nwksc" Oct 07 15:37:27 crc kubenswrapper[4959]: I1007 15:37:27.143857 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nwksc"] Oct 07 15:37:27 crc kubenswrapper[4959]: I1007 15:37:27.173848 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nwksc"] Oct 07 15:37:27 crc kubenswrapper[4959]: I1007 15:37:27.226713 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-msjqn_ddaee6a0-3ed4-4d4f-80a8-686b53ce5a94/openstack-network-exporter/0.log" Oct 07 15:37:27 crc kubenswrapper[4959]: I1007 15:37:27.662694 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v6cb5_2d0933d8-59fe-4293-998c-8a236af482e1/ovsdb-server-init/0.log" Oct 07 15:37:28 crc kubenswrapper[4959]: I1007 15:37:28.012042 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v6cb5_2d0933d8-59fe-4293-998c-8a236af482e1/ovsdb-server-init/0.log" Oct 07 15:37:28 crc kubenswrapper[4959]: I1007 15:37:28.022031 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v6cb5_2d0933d8-59fe-4293-998c-8a236af482e1/ovs-vswitchd/0.log" Oct 07 15:37:28 crc kubenswrapper[4959]: I1007 15:37:28.386775 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v6cb5_2d0933d8-59fe-4293-998c-8a236af482e1/ovsdb-server/0.log" Oct 07 15:37:28 crc kubenswrapper[4959]: I1007 15:37:28.522626 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_a046abc0-ec2e-42ac-ad49-e336824e9317/nova-metadata-metadata/0.log" Oct 07 15:37:28 crc kubenswrapper[4959]: I1007 15:37:28.675805 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fad6288b-9766-448d-9f91-374bc79257dd" path="/var/lib/kubelet/pods/fad6288b-9766-448d-9f91-374bc79257dd/volumes" Oct 07 15:37:28 crc kubenswrapper[4959]: I1007 15:37:28.899487 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-q5v9v_603a8fc2-2eac-47b7-8e97-cd117033a313/ovn-controller/0.log" Oct 07 15:37:29 crc kubenswrapper[4959]: I1007 15:37:29.132683 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-lt62n_64d9894c-3fa2-4dfb-85c7-c8596268ff6b/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 15:37:29 crc kubenswrapper[4959]: I1007 15:37:29.785018 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c76f0f7e-182d-40e0-b4bd-34076ba831e3/ovn-northd/0.log" Oct 07 15:37:29 crc kubenswrapper[4959]: I1007 15:37:29.810049 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c76f0f7e-182d-40e0-b4bd-34076ba831e3/openstack-network-exporter/0.log" Oct 07 15:37:30 crc kubenswrapper[4959]: I1007 15:37:30.158484 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ac9cc42f-0fad-48ad-b4b2-3494a5d10a61/openstack-network-exporter/0.log" Oct 07 15:37:30 crc kubenswrapper[4959]: I1007 15:37:30.181788 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_938d13f8-ccf7-4723-a7ff-3e16f7571eac/memcached/0.log" Oct 07 15:37:30 crc kubenswrapper[4959]: I1007 15:37:30.280650 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ac9cc42f-0fad-48ad-b4b2-3494a5d10a61/ovsdbserver-nb/0.log" Oct 07 15:37:30 crc kubenswrapper[4959]: I1007 15:37:30.595196 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_454a13c2-ada4-4885-89b2-efdde4705730/ovsdbserver-sb/0.log" Oct 07 15:37:30 crc kubenswrapper[4959]: I1007 15:37:30.599408 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_454a13c2-ada4-4885-89b2-efdde4705730/openstack-network-exporter/0.log" Oct 07 15:37:31 crc kubenswrapper[4959]: I1007 15:37:31.029658 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0265cd4e-529d-40a6-a14c-6f39d6068633/setup-container/0.log" Oct 07 15:37:31 crc kubenswrapper[4959]: I1007 15:37:31.236404 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5c4946988b-h259p_342978a1-dc32-4347-bf34-6782c201f033/placement-api/0.log" Oct 07 15:37:31 crc kubenswrapper[4959]: I1007 15:37:31.321116 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5c4946988b-h259p_342978a1-dc32-4347-bf34-6782c201f033/placement-log/0.log" Oct 07 15:37:31 crc kubenswrapper[4959]: I1007 15:37:31.445312 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0265cd4e-529d-40a6-a14c-6f39d6068633/setup-container/0.log" Oct 07 15:37:31 crc kubenswrapper[4959]: I1007 15:37:31.584469 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0265cd4e-529d-40a6-a14c-6f39d6068633/rabbitmq/0.log" Oct 07 15:37:31 crc kubenswrapper[4959]: I1007 15:37:31.650717 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_439982d5-b039-45c1-9300-f39ce049cf33/setup-container/0.log" Oct 07 15:37:31 crc kubenswrapper[4959]: I1007 15:37:31.867828 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_439982d5-b039-45c1-9300-f39ce049cf33/setup-container/0.log" Oct 07 15:37:31 crc kubenswrapper[4959]: I1007 15:37:31.915817 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_439982d5-b039-45c1-9300-f39ce049cf33/rabbitmq/0.log" Oct 07 15:37:31 crc kubenswrapper[4959]: I1007 15:37:31.996696 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-v9zrg_00ed94a9-b896-493f-b1bd-8e161a45139c/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 15:37:32 crc kubenswrapper[4959]: I1007 15:37:32.190814 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-v692z_4064e315-7d27-4e7a-b756-a5f01d7db24a/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 15:37:32 crc kubenswrapper[4959]: I1007 15:37:32.289263 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-s8dcw_0e2eb772-e3a8-427f-a226-7aa85dd69238/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 15:37:32 crc kubenswrapper[4959]: I1007 15:37:32.472012 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-ms8vj_634f1665-e210-4c71-9c7a-8cf30b5f3fec/ssh-known-hosts-edpm-deployment/0.log" Oct 07 15:37:32 crc kubenswrapper[4959]: I1007 15:37:32.914276 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest-s01-single-test_bc180a96-fac6-4b97-97bf-11b7c7ceff8a/tempest-tests-tempest-tests-runner/0.log" Oct 07 15:37:33 crc kubenswrapper[4959]: I1007 15:37:33.153806 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-ansibletest-ansibletest-ansibletest_409bd790-46de-4dc6-bfae-6a75d5117a74/test-operator-logs-container/0.log" Oct 07 15:37:33 crc kubenswrapper[4959]: I1007 15:37:33.233477 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-horizontest-horizontest-tests-horizontest_f51247e0-f093-4d23-83a8-d8713b783834/test-operator-logs-container/0.log" Oct 07 15:37:33 crc kubenswrapper[4959]: I1007 15:37:33.495153 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_5eb07cef-84bc-4028-afd3-b88f5ee8ef75/test-operator-logs-container/0.log" Oct 07 15:37:33 crc kubenswrapper[4959]: I1007 15:37:33.591317 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest-s00-full_b68f21fe-48e8-4b78-8476-ec53a2ca30c8/tempest-tests-tempest-tests-runner/0.log" Oct 07 15:37:33 crc kubenswrapper[4959]: I1007 15:37:33.738787 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tobiko-tobiko-tests-tobiko_e58ced67-303b-467a-b929-7de1e89d2317/test-operator-logs-container/0.log" Oct 07 15:37:33 crc kubenswrapper[4959]: I1007 15:37:33.826733 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tobiko-tests-tobiko-s00-podified-functional_a646a7f2-705e-4278-9950-1a20b2db0149/tobiko-tests-tobiko/0.log" Oct 07 15:37:34 crc kubenswrapper[4959]: I1007 15:37:34.006213 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tobiko-tests-tobiko-s01-sanity_11e9b8da-1cca-4958-8838-4e7fc1364a78/tobiko-tests-tobiko/0.log" Oct 07 15:37:34 crc kubenswrapper[4959]: I1007 15:37:34.103987 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-zblfq_f81e9c7e-9754-410a-9ac3-df9f6fcf20e9/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 07 15:37:36 crc kubenswrapper[4959]: I1007 15:37:36.720992 4959 scope.go:117] "RemoveContainer" containerID="6424b528e2636d9fd5db8a284452d6d3698a0710c9d3984ee3447099dbf624a2" Oct 07 15:37:36 crc kubenswrapper[4959]: I1007 15:37:36.766373 4959 scope.go:117] "RemoveContainer" containerID="1360d2e1c0977ca6158531cc5726f0bfb4a0be4c46d55bf75db0e96c89ce2f4b" Oct 07 15:37:36 crc kubenswrapper[4959]: I1007 15:37:36.802191 4959 scope.go:117] "RemoveContainer" containerID="7c28e4501f6143d9a431c9062d327051d5d1eb10a33bab38661542680c104d28" Oct 07 15:37:40 crc kubenswrapper[4959]: I1007 15:37:40.653529 4959 scope.go:117] "RemoveContainer" containerID="02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a" Oct 07 15:37:40 crc kubenswrapper[4959]: E1007 15:37:40.654448 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:37:52 crc kubenswrapper[4959]: I1007 15:37:52.653896 4959 scope.go:117] "RemoveContainer" containerID="02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a" Oct 07 15:37:53 crc kubenswrapper[4959]: I1007 15:37:53.411608 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerStarted","Data":"6dd605c45ef00573c72aea18d5f4874a4f19110dd7cbf9ba997a985a527c37db"} Oct 07 15:38:00 crc kubenswrapper[4959]: E1007 15:38:00.654013 4959 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 07 15:39:03 crc kubenswrapper[4959]: I1007 15:39:03.094398 4959 generic.go:334] "Generic (PLEG): container finished" podID="7a76f0ab-a8cf-4470-bc5f-09e711ba3fee" containerID="e2239caf5f94b699aa8d89234db94aa42555f219249ef40a61f0d89eab4dcf1b" exitCode=0 Oct 07 15:39:03 crc kubenswrapper[4959]: I1007 15:39:03.094478 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tlcvp/crc-debug-2c9dg" event={"ID":"7a76f0ab-a8cf-4470-bc5f-09e711ba3fee","Type":"ContainerDied","Data":"e2239caf5f94b699aa8d89234db94aa42555f219249ef40a61f0d89eab4dcf1b"} Oct 07 15:39:04 crc kubenswrapper[4959]: I1007 15:39:04.218977 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tlcvp/crc-debug-2c9dg" Oct 07 15:39:04 crc kubenswrapper[4959]: I1007 15:39:04.253531 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-tlcvp/crc-debug-2c9dg"] Oct 07 15:39:04 crc kubenswrapper[4959]: I1007 15:39:04.262884 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-tlcvp/crc-debug-2c9dg"] Oct 07 15:39:04 crc kubenswrapper[4959]: I1007 15:39:04.332078 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zth8w\" (UniqueName: \"kubernetes.io/projected/7a76f0ab-a8cf-4470-bc5f-09e711ba3fee-kube-api-access-zth8w\") pod \"7a76f0ab-a8cf-4470-bc5f-09e711ba3fee\" (UID: \"7a76f0ab-a8cf-4470-bc5f-09e711ba3fee\") " Oct 07 15:39:04 crc kubenswrapper[4959]: I1007 15:39:04.332284 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7a76f0ab-a8cf-4470-bc5f-09e711ba3fee-host\") pod \"7a76f0ab-a8cf-4470-bc5f-09e711ba3fee\" (UID: \"7a76f0ab-a8cf-4470-bc5f-09e711ba3fee\") " Oct 07 15:39:04 crc kubenswrapper[4959]: I1007 15:39:04.332659 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a76f0ab-a8cf-4470-bc5f-09e711ba3fee-host" (OuterVolumeSpecName: "host") pod "7a76f0ab-a8cf-4470-bc5f-09e711ba3fee" (UID: "7a76f0ab-a8cf-4470-bc5f-09e711ba3fee"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 15:39:04 crc kubenswrapper[4959]: I1007 15:39:04.333294 4959 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7a76f0ab-a8cf-4470-bc5f-09e711ba3fee-host\") on node \"crc\" DevicePath \"\"" Oct 07 15:39:04 crc kubenswrapper[4959]: I1007 15:39:04.337587 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a76f0ab-a8cf-4470-bc5f-09e711ba3fee-kube-api-access-zth8w" (OuterVolumeSpecName: "kube-api-access-zth8w") pod "7a76f0ab-a8cf-4470-bc5f-09e711ba3fee" (UID: "7a76f0ab-a8cf-4470-bc5f-09e711ba3fee"). InnerVolumeSpecName "kube-api-access-zth8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 15:39:04 crc kubenswrapper[4959]: I1007 15:39:04.435248 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zth8w\" (UniqueName: \"kubernetes.io/projected/7a76f0ab-a8cf-4470-bc5f-09e711ba3fee-kube-api-access-zth8w\") on node \"crc\" DevicePath \"\"" Oct 07 15:39:04 crc kubenswrapper[4959]: I1007 15:39:04.674471 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a76f0ab-a8cf-4470-bc5f-09e711ba3fee" path="/var/lib/kubelet/pods/7a76f0ab-a8cf-4470-bc5f-09e711ba3fee/volumes" Oct 07 15:39:05 crc kubenswrapper[4959]: I1007 15:39:05.117868 4959 scope.go:117] "RemoveContainer" containerID="e2239caf5f94b699aa8d89234db94aa42555f219249ef40a61f0d89eab4dcf1b" Oct 07 15:39:05 crc kubenswrapper[4959]: I1007 15:39:05.118533 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tlcvp/crc-debug-2c9dg" Oct 07 15:39:05 crc kubenswrapper[4959]: I1007 15:39:05.490845 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-tlcvp/crc-debug-rwvwk"] Oct 07 15:39:05 crc kubenswrapper[4959]: E1007 15:39:05.491526 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a76f0ab-a8cf-4470-bc5f-09e711ba3fee" containerName="container-00" Oct 07 15:39:05 crc kubenswrapper[4959]: I1007 15:39:05.491548 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a76f0ab-a8cf-4470-bc5f-09e711ba3fee" containerName="container-00" Oct 07 15:39:05 crc kubenswrapper[4959]: E1007 15:39:05.491580 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce3552bd-baa1-4b76-9943-8cc375c11945" containerName="extract-content" Oct 07 15:39:05 crc kubenswrapper[4959]: I1007 15:39:05.491599 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce3552bd-baa1-4b76-9943-8cc375c11945" containerName="extract-content" Oct 07 15:39:05 crc kubenswrapper[4959]: E1007 15:39:05.491622 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce3552bd-baa1-4b76-9943-8cc375c11945" containerName="extract-utilities" Oct 07 15:39:05 crc kubenswrapper[4959]: I1007 15:39:05.491635 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce3552bd-baa1-4b76-9943-8cc375c11945" containerName="extract-utilities" Oct 07 15:39:05 crc kubenswrapper[4959]: E1007 15:39:05.491651 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce3552bd-baa1-4b76-9943-8cc375c11945" containerName="registry-server" Oct 07 15:39:05 crc kubenswrapper[4959]: I1007 15:39:05.491659 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce3552bd-baa1-4b76-9943-8cc375c11945" containerName="registry-server" Oct 07 15:39:05 crc kubenswrapper[4959]: E1007 15:39:05.491669 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fad6288b-9766-448d-9f91-374bc79257dd" containerName="extract-utilities" Oct 07 15:39:05 crc kubenswrapper[4959]: I1007 15:39:05.491676 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="fad6288b-9766-448d-9f91-374bc79257dd" containerName="extract-utilities" Oct 07 15:39:05 crc kubenswrapper[4959]: E1007 15:39:05.491690 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fad6288b-9766-448d-9f91-374bc79257dd" containerName="registry-server" Oct 07 15:39:05 crc kubenswrapper[4959]: I1007 15:39:05.491697 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="fad6288b-9766-448d-9f91-374bc79257dd" containerName="registry-server" Oct 07 15:39:05 crc kubenswrapper[4959]: E1007 15:39:05.491741 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fad6288b-9766-448d-9f91-374bc79257dd" containerName="extract-content" Oct 07 15:39:05 crc kubenswrapper[4959]: I1007 15:39:05.491749 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="fad6288b-9766-448d-9f91-374bc79257dd" containerName="extract-content" Oct 07 15:39:05 crc kubenswrapper[4959]: I1007 15:39:05.492010 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="fad6288b-9766-448d-9f91-374bc79257dd" containerName="registry-server" Oct 07 15:39:05 crc kubenswrapper[4959]: I1007 15:39:05.492037 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a76f0ab-a8cf-4470-bc5f-09e711ba3fee" containerName="container-00" Oct 07 15:39:05 crc kubenswrapper[4959]: I1007 15:39:05.492055 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce3552bd-baa1-4b76-9943-8cc375c11945" containerName="registry-server" Oct 07 15:39:05 crc kubenswrapper[4959]: I1007 15:39:05.493281 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tlcvp/crc-debug-rwvwk" Oct 07 15:39:05 crc kubenswrapper[4959]: I1007 15:39:05.560988 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/08c19d97-f925-4f1b-8639-2a71e078c312-host\") pod \"crc-debug-rwvwk\" (UID: \"08c19d97-f925-4f1b-8639-2a71e078c312\") " pod="openshift-must-gather-tlcvp/crc-debug-rwvwk" Oct 07 15:39:05 crc kubenswrapper[4959]: I1007 15:39:05.561208 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25v2v\" (UniqueName: \"kubernetes.io/projected/08c19d97-f925-4f1b-8639-2a71e078c312-kube-api-access-25v2v\") pod \"crc-debug-rwvwk\" (UID: \"08c19d97-f925-4f1b-8639-2a71e078c312\") " pod="openshift-must-gather-tlcvp/crc-debug-rwvwk" Oct 07 15:39:05 crc kubenswrapper[4959]: I1007 15:39:05.663385 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25v2v\" (UniqueName: \"kubernetes.io/projected/08c19d97-f925-4f1b-8639-2a71e078c312-kube-api-access-25v2v\") pod \"crc-debug-rwvwk\" (UID: \"08c19d97-f925-4f1b-8639-2a71e078c312\") " pod="openshift-must-gather-tlcvp/crc-debug-rwvwk" Oct 07 15:39:05 crc kubenswrapper[4959]: I1007 15:39:05.663501 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/08c19d97-f925-4f1b-8639-2a71e078c312-host\") pod \"crc-debug-rwvwk\" (UID: \"08c19d97-f925-4f1b-8639-2a71e078c312\") " pod="openshift-must-gather-tlcvp/crc-debug-rwvwk" Oct 07 15:39:05 crc kubenswrapper[4959]: I1007 15:39:05.663689 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/08c19d97-f925-4f1b-8639-2a71e078c312-host\") pod \"crc-debug-rwvwk\" (UID: \"08c19d97-f925-4f1b-8639-2a71e078c312\") " pod="openshift-must-gather-tlcvp/crc-debug-rwvwk" Oct 07 15:39:05 crc kubenswrapper[4959]: I1007 15:39:05.682617 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25v2v\" (UniqueName: \"kubernetes.io/projected/08c19d97-f925-4f1b-8639-2a71e078c312-kube-api-access-25v2v\") pod \"crc-debug-rwvwk\" (UID: \"08c19d97-f925-4f1b-8639-2a71e078c312\") " pod="openshift-must-gather-tlcvp/crc-debug-rwvwk" Oct 07 15:39:05 crc kubenswrapper[4959]: I1007 15:39:05.816805 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tlcvp/crc-debug-rwvwk" Oct 07 15:39:05 crc kubenswrapper[4959]: W1007 15:39:05.858824 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08c19d97_f925_4f1b_8639_2a71e078c312.slice/crio-2dab4fef846bededf78d6b9a52d6a4fa411096450dc00565d5169c9d0a9ec025 WatchSource:0}: Error finding container 2dab4fef846bededf78d6b9a52d6a4fa411096450dc00565d5169c9d0a9ec025: Status 404 returned error can't find the container with id 2dab4fef846bededf78d6b9a52d6a4fa411096450dc00565d5169c9d0a9ec025 Oct 07 15:39:06 crc kubenswrapper[4959]: I1007 15:39:06.131491 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tlcvp/crc-debug-rwvwk" event={"ID":"08c19d97-f925-4f1b-8639-2a71e078c312","Type":"ContainerStarted","Data":"2dab4fef846bededf78d6b9a52d6a4fa411096450dc00565d5169c9d0a9ec025"} Oct 07 15:39:07 crc kubenswrapper[4959]: I1007 15:39:07.142215 4959 generic.go:334] "Generic (PLEG): container finished" podID="08c19d97-f925-4f1b-8639-2a71e078c312" containerID="d9ff643ce7ebe5d9592f4e04104a93b0cc3a3afc9298f8db3a8eedd592a8330d" exitCode=0 Oct 07 15:39:07 crc kubenswrapper[4959]: I1007 15:39:07.142323 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tlcvp/crc-debug-rwvwk" event={"ID":"08c19d97-f925-4f1b-8639-2a71e078c312","Type":"ContainerDied","Data":"d9ff643ce7ebe5d9592f4e04104a93b0cc3a3afc9298f8db3a8eedd592a8330d"} Oct 07 15:39:08 crc kubenswrapper[4959]: I1007 15:39:08.318452 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tlcvp/crc-debug-rwvwk" Oct 07 15:39:08 crc kubenswrapper[4959]: I1007 15:39:08.427982 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25v2v\" (UniqueName: \"kubernetes.io/projected/08c19d97-f925-4f1b-8639-2a71e078c312-kube-api-access-25v2v\") pod \"08c19d97-f925-4f1b-8639-2a71e078c312\" (UID: \"08c19d97-f925-4f1b-8639-2a71e078c312\") " Oct 07 15:39:08 crc kubenswrapper[4959]: I1007 15:39:08.428054 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/08c19d97-f925-4f1b-8639-2a71e078c312-host\") pod \"08c19d97-f925-4f1b-8639-2a71e078c312\" (UID: \"08c19d97-f925-4f1b-8639-2a71e078c312\") " Oct 07 15:39:08 crc kubenswrapper[4959]: I1007 15:39:08.429297 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/08c19d97-f925-4f1b-8639-2a71e078c312-host" (OuterVolumeSpecName: "host") pod "08c19d97-f925-4f1b-8639-2a71e078c312" (UID: "08c19d97-f925-4f1b-8639-2a71e078c312"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 15:39:08 crc kubenswrapper[4959]: I1007 15:39:08.431396 4959 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/08c19d97-f925-4f1b-8639-2a71e078c312-host\") on node \"crc\" DevicePath \"\"" Oct 07 15:39:08 crc kubenswrapper[4959]: I1007 15:39:08.444089 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08c19d97-f925-4f1b-8639-2a71e078c312-kube-api-access-25v2v" (OuterVolumeSpecName: "kube-api-access-25v2v") pod "08c19d97-f925-4f1b-8639-2a71e078c312" (UID: "08c19d97-f925-4f1b-8639-2a71e078c312"). InnerVolumeSpecName "kube-api-access-25v2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 15:39:08 crc kubenswrapper[4959]: I1007 15:39:08.533688 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25v2v\" (UniqueName: \"kubernetes.io/projected/08c19d97-f925-4f1b-8639-2a71e078c312-kube-api-access-25v2v\") on node \"crc\" DevicePath \"\"" Oct 07 15:39:09 crc kubenswrapper[4959]: I1007 15:39:09.166457 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tlcvp/crc-debug-rwvwk" event={"ID":"08c19d97-f925-4f1b-8639-2a71e078c312","Type":"ContainerDied","Data":"2dab4fef846bededf78d6b9a52d6a4fa411096450dc00565d5169c9d0a9ec025"} Oct 07 15:39:09 crc kubenswrapper[4959]: I1007 15:39:09.166507 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2dab4fef846bededf78d6b9a52d6a4fa411096450dc00565d5169c9d0a9ec025" Oct 07 15:39:09 crc kubenswrapper[4959]: I1007 15:39:09.166516 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tlcvp/crc-debug-rwvwk" Oct 07 15:39:14 crc kubenswrapper[4959]: E1007 15:39:14.654289 4959 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 07 15:39:18 crc kubenswrapper[4959]: I1007 15:39:18.201519 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-tlcvp/crc-debug-rwvwk"] Oct 07 15:39:18 crc kubenswrapper[4959]: I1007 15:39:18.220383 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-tlcvp/crc-debug-rwvwk"] Oct 07 15:39:18 crc kubenswrapper[4959]: I1007 15:39:18.667072 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08c19d97-f925-4f1b-8639-2a71e078c312" path="/var/lib/kubelet/pods/08c19d97-f925-4f1b-8639-2a71e078c312/volumes" Oct 07 15:39:19 crc kubenswrapper[4959]: I1007 15:39:19.366735 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-tlcvp/crc-debug-f4r4p"] Oct 07 15:39:19 crc kubenswrapper[4959]: E1007 15:39:19.367439 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08c19d97-f925-4f1b-8639-2a71e078c312" containerName="container-00" Oct 07 15:39:19 crc kubenswrapper[4959]: I1007 15:39:19.367459 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="08c19d97-f925-4f1b-8639-2a71e078c312" containerName="container-00" Oct 07 15:39:19 crc kubenswrapper[4959]: I1007 15:39:19.367675 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="08c19d97-f925-4f1b-8639-2a71e078c312" containerName="container-00" Oct 07 15:39:19 crc kubenswrapper[4959]: I1007 15:39:19.368314 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tlcvp/crc-debug-f4r4p" Oct 07 15:39:19 crc kubenswrapper[4959]: I1007 15:39:19.446238 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dabfbd4d-255f-45f7-ab12-6450a2d723b2-host\") pod \"crc-debug-f4r4p\" (UID: \"dabfbd4d-255f-45f7-ab12-6450a2d723b2\") " pod="openshift-must-gather-tlcvp/crc-debug-f4r4p" Oct 07 15:39:19 crc kubenswrapper[4959]: I1007 15:39:19.446324 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmdgb\" (UniqueName: \"kubernetes.io/projected/dabfbd4d-255f-45f7-ab12-6450a2d723b2-kube-api-access-vmdgb\") pod \"crc-debug-f4r4p\" (UID: \"dabfbd4d-255f-45f7-ab12-6450a2d723b2\") " pod="openshift-must-gather-tlcvp/crc-debug-f4r4p" Oct 07 15:39:19 crc kubenswrapper[4959]: I1007 15:39:19.548528 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dabfbd4d-255f-45f7-ab12-6450a2d723b2-host\") pod \"crc-debug-f4r4p\" (UID: \"dabfbd4d-255f-45f7-ab12-6450a2d723b2\") " pod="openshift-must-gather-tlcvp/crc-debug-f4r4p" Oct 07 15:39:19 crc kubenswrapper[4959]: I1007 15:39:19.548630 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmdgb\" (UniqueName: \"kubernetes.io/projected/dabfbd4d-255f-45f7-ab12-6450a2d723b2-kube-api-access-vmdgb\") pod \"crc-debug-f4r4p\" (UID: \"dabfbd4d-255f-45f7-ab12-6450a2d723b2\") " pod="openshift-must-gather-tlcvp/crc-debug-f4r4p" Oct 07 15:39:19 crc kubenswrapper[4959]: I1007 15:39:19.548751 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dabfbd4d-255f-45f7-ab12-6450a2d723b2-host\") pod \"crc-debug-f4r4p\" (UID: \"dabfbd4d-255f-45f7-ab12-6450a2d723b2\") " pod="openshift-must-gather-tlcvp/crc-debug-f4r4p" Oct 07 15:39:19 crc kubenswrapper[4959]: I1007 15:39:19.572897 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmdgb\" (UniqueName: \"kubernetes.io/projected/dabfbd4d-255f-45f7-ab12-6450a2d723b2-kube-api-access-vmdgb\") pod \"crc-debug-f4r4p\" (UID: \"dabfbd4d-255f-45f7-ab12-6450a2d723b2\") " pod="openshift-must-gather-tlcvp/crc-debug-f4r4p" Oct 07 15:39:19 crc kubenswrapper[4959]: I1007 15:39:19.685573 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tlcvp/crc-debug-f4r4p" Oct 07 15:39:20 crc kubenswrapper[4959]: I1007 15:39:20.270114 4959 generic.go:334] "Generic (PLEG): container finished" podID="dabfbd4d-255f-45f7-ab12-6450a2d723b2" containerID="ff01e4a07772711cc61f48c3a36e5cd89661d971ac366521e423fbc0424d03ad" exitCode=0 Oct 07 15:39:20 crc kubenswrapper[4959]: I1007 15:39:20.270200 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tlcvp/crc-debug-f4r4p" event={"ID":"dabfbd4d-255f-45f7-ab12-6450a2d723b2","Type":"ContainerDied","Data":"ff01e4a07772711cc61f48c3a36e5cd89661d971ac366521e423fbc0424d03ad"} Oct 07 15:39:20 crc kubenswrapper[4959]: I1007 15:39:20.270643 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tlcvp/crc-debug-f4r4p" event={"ID":"dabfbd4d-255f-45f7-ab12-6450a2d723b2","Type":"ContainerStarted","Data":"27ccb80a9fcbd3094bc3949a1dc85a39f30c122615d9a1d9631ac50d31fd98a2"} Oct 07 15:39:20 crc kubenswrapper[4959]: I1007 15:39:20.305980 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-tlcvp/crc-debug-f4r4p"] Oct 07 15:39:20 crc kubenswrapper[4959]: I1007 15:39:20.314669 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-tlcvp/crc-debug-f4r4p"] Oct 07 15:39:21 crc kubenswrapper[4959]: I1007 15:39:21.422208 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tlcvp/crc-debug-f4r4p" Oct 07 15:39:21 crc kubenswrapper[4959]: I1007 15:39:21.497235 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmdgb\" (UniqueName: \"kubernetes.io/projected/dabfbd4d-255f-45f7-ab12-6450a2d723b2-kube-api-access-vmdgb\") pod \"dabfbd4d-255f-45f7-ab12-6450a2d723b2\" (UID: \"dabfbd4d-255f-45f7-ab12-6450a2d723b2\") " Oct 07 15:39:21 crc kubenswrapper[4959]: I1007 15:39:21.497638 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dabfbd4d-255f-45f7-ab12-6450a2d723b2-host\") pod \"dabfbd4d-255f-45f7-ab12-6450a2d723b2\" (UID: \"dabfbd4d-255f-45f7-ab12-6450a2d723b2\") " Oct 07 15:39:21 crc kubenswrapper[4959]: I1007 15:39:21.497734 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dabfbd4d-255f-45f7-ab12-6450a2d723b2-host" (OuterVolumeSpecName: "host") pod "dabfbd4d-255f-45f7-ab12-6450a2d723b2" (UID: "dabfbd4d-255f-45f7-ab12-6450a2d723b2"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 07 15:39:21 crc kubenswrapper[4959]: I1007 15:39:21.498519 4959 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dabfbd4d-255f-45f7-ab12-6450a2d723b2-host\") on node \"crc\" DevicePath \"\"" Oct 07 15:39:21 crc kubenswrapper[4959]: I1007 15:39:21.513430 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dabfbd4d-255f-45f7-ab12-6450a2d723b2-kube-api-access-vmdgb" (OuterVolumeSpecName: "kube-api-access-vmdgb") pod "dabfbd4d-255f-45f7-ab12-6450a2d723b2" (UID: "dabfbd4d-255f-45f7-ab12-6450a2d723b2"). InnerVolumeSpecName "kube-api-access-vmdgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 15:39:21 crc kubenswrapper[4959]: I1007 15:39:21.601126 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmdgb\" (UniqueName: \"kubernetes.io/projected/dabfbd4d-255f-45f7-ab12-6450a2d723b2-kube-api-access-vmdgb\") on node \"crc\" DevicePath \"\"" Oct 07 15:39:22 crc kubenswrapper[4959]: I1007 15:39:22.075808 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv_209faffc-bcb0-4a9b-8558-eab623e1b675/util/0.log" Oct 07 15:39:22 crc kubenswrapper[4959]: I1007 15:39:22.178511 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv_209faffc-bcb0-4a9b-8558-eab623e1b675/util/0.log" Oct 07 15:39:22 crc kubenswrapper[4959]: I1007 15:39:22.265701 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv_209faffc-bcb0-4a9b-8558-eab623e1b675/pull/0.log" Oct 07 15:39:22 crc kubenswrapper[4959]: I1007 15:39:22.270400 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv_209faffc-bcb0-4a9b-8558-eab623e1b675/pull/0.log" Oct 07 15:39:22 crc kubenswrapper[4959]: I1007 15:39:22.296881 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tlcvp/crc-debug-f4r4p" Oct 07 15:39:22 crc kubenswrapper[4959]: I1007 15:39:22.296949 4959 scope.go:117] "RemoveContainer" containerID="ff01e4a07772711cc61f48c3a36e5cd89661d971ac366521e423fbc0424d03ad" Oct 07 15:39:22 crc kubenswrapper[4959]: I1007 15:39:22.495834 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv_209faffc-bcb0-4a9b-8558-eab623e1b675/pull/0.log" Oct 07 15:39:22 crc kubenswrapper[4959]: I1007 15:39:22.533829 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv_209faffc-bcb0-4a9b-8558-eab623e1b675/util/0.log" Oct 07 15:39:22 crc kubenswrapper[4959]: I1007 15:39:22.555598 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_563473821d46f325f62a2983c417e07a5da08327e4dc09e802caf673b0hg8bv_209faffc-bcb0-4a9b-8558-eab623e1b675/extract/0.log" Oct 07 15:39:22 crc kubenswrapper[4959]: I1007 15:39:22.666066 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dabfbd4d-255f-45f7-ab12-6450a2d723b2" path="/var/lib/kubelet/pods/dabfbd4d-255f-45f7-ab12-6450a2d723b2/volumes" Oct 07 15:39:22 crc kubenswrapper[4959]: I1007 15:39:22.741195 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f56ff694-9k7x7_1e3c7b8e-8a44-4539-9e01-4fe9ce17266e/kube-rbac-proxy/0.log" Oct 07 15:39:22 crc kubenswrapper[4959]: I1007 15:39:22.763399 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-84bd8f6848-85vgl_2b84227a-34a9-4969-97be-4750eb555f22/kube-rbac-proxy/0.log" Oct 07 15:39:22 crc kubenswrapper[4959]: I1007 15:39:22.776686 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f56ff694-9k7x7_1e3c7b8e-8a44-4539-9e01-4fe9ce17266e/manager/0.log" Oct 07 15:39:22 crc kubenswrapper[4959]: I1007 15:39:22.985917 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-84bd8f6848-85vgl_2b84227a-34a9-4969-97be-4750eb555f22/manager/0.log" Oct 07 15:39:22 crc kubenswrapper[4959]: I1007 15:39:22.989328 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-58d86cd59d-pnbwt_e46e007a-5053-443a-9619-247dd3d79a8e/kube-rbac-proxy/0.log" Oct 07 15:39:23 crc kubenswrapper[4959]: I1007 15:39:23.057604 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-58d86cd59d-pnbwt_e46e007a-5053-443a-9619-247dd3d79a8e/manager/0.log" Oct 07 15:39:23 crc kubenswrapper[4959]: I1007 15:39:23.191506 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-fd648f65-bc7rh_ca1a8a8b-1649-49b3-b595-0e12408673c9/kube-rbac-proxy/0.log" Oct 07 15:39:23 crc kubenswrapper[4959]: I1007 15:39:23.331637 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-fd648f65-bc7rh_ca1a8a8b-1649-49b3-b595-0e12408673c9/manager/0.log" Oct 07 15:39:23 crc kubenswrapper[4959]: I1007 15:39:23.406396 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-7ccfc8cf49-q56bw_98825e89-a8c2-4023-b139-4c6f5d9b7987/kube-rbac-proxy/0.log" Oct 07 15:39:23 crc kubenswrapper[4959]: I1007 15:39:23.537913 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-7ccfc8cf49-q56bw_98825e89-a8c2-4023-b139-4c6f5d9b7987/manager/0.log" Oct 07 15:39:23 crc kubenswrapper[4959]: I1007 15:39:23.589218 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5b477879bc-p6mgb_c21b1db6-de7b-48eb-9796-b7d45a638e8c/kube-rbac-proxy/0.log" Oct 07 15:39:23 crc kubenswrapper[4959]: I1007 15:39:23.643504 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5b477879bc-p6mgb_c21b1db6-de7b-48eb-9796-b7d45a638e8c/manager/0.log" Oct 07 15:39:23 crc kubenswrapper[4959]: I1007 15:39:23.811506 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-84788b6bc5-rdj2q_77cf7bcc-63b9-4ee3-84bb-4d4b784750c6/kube-rbac-proxy/0.log" Oct 07 15:39:24 crc kubenswrapper[4959]: I1007 15:39:24.038783 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-84788b6bc5-rdj2q_77cf7bcc-63b9-4ee3-84bb-4d4b784750c6/manager/0.log" Oct 07 15:39:24 crc kubenswrapper[4959]: I1007 15:39:24.044440 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5467f8988c-l4hdq_c47e0f74-de21-41f6-9142-85d47fcafc60/manager/0.log" Oct 07 15:39:24 crc kubenswrapper[4959]: I1007 15:39:24.054248 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5467f8988c-l4hdq_c47e0f74-de21-41f6-9142-85d47fcafc60/kube-rbac-proxy/0.log" Oct 07 15:39:24 crc kubenswrapper[4959]: I1007 15:39:24.249939 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5b84cc7657-k5rmd_64565126-cba9-4da4-8e0e-7a75f2acc4f1/kube-rbac-proxy/0.log" Oct 07 15:39:24 crc kubenswrapper[4959]: I1007 15:39:24.343176 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5b84cc7657-k5rmd_64565126-cba9-4da4-8e0e-7a75f2acc4f1/manager/0.log" Oct 07 15:39:24 crc kubenswrapper[4959]: I1007 15:39:24.469152 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7cb48dbc-hhbqh_801942fb-8de2-449b-9a14-4616bdec95aa/kube-rbac-proxy/0.log" Oct 07 15:39:24 crc kubenswrapper[4959]: I1007 15:39:24.498343 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7cb48dbc-hhbqh_801942fb-8de2-449b-9a14-4616bdec95aa/manager/0.log" Oct 07 15:39:24 crc kubenswrapper[4959]: I1007 15:39:24.568387 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-d6c9dc5bc-ksjkq_e1da143c-64e7-4c75-97fb-4e3553fc2906/kube-rbac-proxy/0.log" Oct 07 15:39:24 crc kubenswrapper[4959]: I1007 15:39:24.722093 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-d6c9dc5bc-ksjkq_e1da143c-64e7-4c75-97fb-4e3553fc2906/manager/0.log" Oct 07 15:39:24 crc kubenswrapper[4959]: I1007 15:39:24.805237 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-69b956fbf6-k4pkf_26ef3538-9145-485b-aa33-f888f7a9d46e/kube-rbac-proxy/0.log" Oct 07 15:39:24 crc kubenswrapper[4959]: I1007 15:39:24.861370 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-69b956fbf6-k4pkf_26ef3538-9145-485b-aa33-f888f7a9d46e/manager/0.log" Oct 07 15:39:25 crc kubenswrapper[4959]: I1007 15:39:25.004446 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-6c9b57c67-mgrc2_97882e07-7ee2-4b7e-86fc-8033bdd78934/kube-rbac-proxy/0.log" Oct 07 15:39:25 crc kubenswrapper[4959]: I1007 15:39:25.162858 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-6c9b57c67-mgrc2_97882e07-7ee2-4b7e-86fc-8033bdd78934/manager/0.log" Oct 07 15:39:25 crc kubenswrapper[4959]: I1007 15:39:25.227574 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69f59f9d8-tfrz7_9bee8f50-89db-4ddb-a033-6f15137aff40/kube-rbac-proxy/0.log" Oct 07 15:39:25 crc kubenswrapper[4959]: I1007 15:39:25.239409 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69f59f9d8-tfrz7_9bee8f50-89db-4ddb-a033-6f15137aff40/manager/0.log" Oct 07 15:39:25 crc kubenswrapper[4959]: I1007 15:39:25.481911 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9_6450acc6-a3fb-429e-903e-8f99ab864a71/kube-rbac-proxy/0.log" Oct 07 15:39:25 crc kubenswrapper[4959]: I1007 15:39:25.498204 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7dcdb4fdb87ljm9_6450acc6-a3fb-429e-903e-8f99ab864a71/manager/0.log" Oct 07 15:39:25 crc kubenswrapper[4959]: I1007 15:39:25.680647 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-fd79fd9-mt8qw_b86a7280-99fd-4b0a-bfeb-cab73d5c360a/kube-rbac-proxy/0.log" Oct 07 15:39:25 crc kubenswrapper[4959]: I1007 15:39:25.810117 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5d4b45574b-njslg_92db920f-830b-4d30-88c6-e9ebb8b83d08/kube-rbac-proxy/0.log" Oct 07 15:39:25 crc kubenswrapper[4959]: I1007 15:39:25.898381 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5d4b45574b-njslg_92db920f-830b-4d30-88c6-e9ebb8b83d08/operator/0.log" Oct 07 15:39:26 crc kubenswrapper[4959]: I1007 15:39:26.128449 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-28476_51c0a7d2-8289-4c41-b419-16db5b361e72/registry-server/0.log" Oct 07 15:39:26 crc kubenswrapper[4959]: I1007 15:39:26.195868 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54d485fd9-7x6df_b6e75c33-53ea-4ecb-bb66-1dfb86201e9d/kube-rbac-proxy/0.log" Oct 07 15:39:26 crc kubenswrapper[4959]: I1007 15:39:26.337224 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54d485fd9-7x6df_b6e75c33-53ea-4ecb-bb66-1dfb86201e9d/manager/0.log" Oct 07 15:39:26 crc kubenswrapper[4959]: I1007 15:39:26.420246 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-66f6d6849b-zdxb9_e1e7bf8b-8f59-486f-920e-adfa13c89d2a/kube-rbac-proxy/0.log" Oct 07 15:39:26 crc kubenswrapper[4959]: I1007 15:39:26.525032 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-66f6d6849b-zdxb9_e1e7bf8b-8f59-486f-920e-adfa13c89d2a/manager/0.log" Oct 07 15:39:26 crc kubenswrapper[4959]: I1007 15:39:26.652848 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-99v5c_54ba157f-2ce3-4fed-aab7-b9e0c1056b78/operator/0.log" Oct 07 15:39:26 crc kubenswrapper[4959]: I1007 15:39:26.792814 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-76d5577b-4tp8k_c28ad232-e3aa-425b-a04d-e1f31cd22770/kube-rbac-proxy/0.log" Oct 07 15:39:27 crc kubenswrapper[4959]: I1007 15:39:27.012175 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-76d5577b-4tp8k_c28ad232-e3aa-425b-a04d-e1f31cd22770/manager/0.log" Oct 07 15:39:27 crc kubenswrapper[4959]: I1007 15:39:27.068469 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-f589c7597-lcjnz_793bfbb8-e13c-4c1c-a490-d422b79b88c5/kube-rbac-proxy/0.log" Oct 07 15:39:27 crc kubenswrapper[4959]: I1007 15:39:27.151796 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-f589c7597-lcjnz_793bfbb8-e13c-4c1c-a490-d422b79b88c5/manager/0.log" Oct 07 15:39:27 crc kubenswrapper[4959]: I1007 15:39:27.223292 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-c4f95597f-8ltkv_9a28faa3-b29b-468f-af7a-3a6f985e7ce9/kube-rbac-proxy/0.log" Oct 07 15:39:27 crc kubenswrapper[4959]: I1007 15:39:27.284564 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-fd79fd9-mt8qw_b86a7280-99fd-4b0a-bfeb-cab73d5c360a/manager/0.log" Oct 07 15:39:27 crc kubenswrapper[4959]: I1007 15:39:27.372645 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-c4f95597f-8ltkv_9a28faa3-b29b-468f-af7a-3a6f985e7ce9/manager/0.log" Oct 07 15:39:27 crc kubenswrapper[4959]: I1007 15:39:27.431709 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5d98cc5575-74gv9_de837fcd-bcf9-4582-86e9-806b5442dae9/kube-rbac-proxy/0.log" Oct 07 15:39:27 crc kubenswrapper[4959]: I1007 15:39:27.508961 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5d98cc5575-74gv9_de837fcd-bcf9-4582-86e9-806b5442dae9/manager/0.log" Oct 07 15:39:43 crc kubenswrapper[4959]: I1007 15:39:43.304985 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-8k48v_5d295e40-2e2f-419e-9db4-cee7392d913e/control-plane-machine-set-operator/0.log" Oct 07 15:39:43 crc kubenswrapper[4959]: I1007 15:39:43.479653 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-dt2k4_1d9c7700-8ba2-444e-a3f0-5c5c6dd05585/kube-rbac-proxy/0.log" Oct 07 15:39:43 crc kubenswrapper[4959]: I1007 15:39:43.511925 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-dt2k4_1d9c7700-8ba2-444e-a3f0-5c5c6dd05585/machine-api-operator/0.log" Oct 07 15:39:55 crc kubenswrapper[4959]: I1007 15:39:55.093378 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-h4qbt_b3b35c9d-c7b3-4baa-a282-464d7598355c/cert-manager-controller/0.log" Oct 07 15:39:55 crc kubenswrapper[4959]: I1007 15:39:55.259564 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-5d4td_a0dc763a-75f9-41bf-bc04-f2da6c5bfc1e/cert-manager-cainjector/0.log" Oct 07 15:39:55 crc kubenswrapper[4959]: I1007 15:39:55.330699 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-pw7vl_0ab413ae-a8d2-41af-8abc-0f4169b3166e/cert-manager-webhook/0.log" Oct 07 15:40:06 crc kubenswrapper[4959]: I1007 15:40:06.508407 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-t7bgt_0d2a66c5-8366-46e5-9c16-ebf1833b0207/nmstate-console-plugin/0.log" Oct 07 15:40:06 crc kubenswrapper[4959]: I1007 15:40:06.690852 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-vsp7j_013ff83c-f506-43cc-877b-dc67c6bb7c10/nmstate-handler/0.log" Oct 07 15:40:06 crc kubenswrapper[4959]: I1007 15:40:06.733014 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-f9jjx_1e34329c-0703-41ea-81b5-cd2056ab1491/nmstate-metrics/0.log" Oct 07 15:40:06 crc kubenswrapper[4959]: I1007 15:40:06.733641 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-f9jjx_1e34329c-0703-41ea-81b5-cd2056ab1491/kube-rbac-proxy/0.log" Oct 07 15:40:06 crc kubenswrapper[4959]: I1007 15:40:06.896300 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-qf9tf_9569c43f-7384-4561-b6fe-231eedaba53a/nmstate-operator/0.log" Oct 07 15:40:06 crc kubenswrapper[4959]: I1007 15:40:06.931563 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-v5vcx_5d7ebbd8-5199-43e0-a92d-54f120ee968a/nmstate-webhook/0.log" Oct 07 15:40:20 crc kubenswrapper[4959]: I1007 15:40:20.491900 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-nbdf7_e2e00fa0-bc56-4a3f-9796-bbc1ec5dc47d/kube-rbac-proxy/0.log" Oct 07 15:40:20 crc kubenswrapper[4959]: I1007 15:40:20.632367 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 15:40:20 crc kubenswrapper[4959]: I1007 15:40:20.632647 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 15:40:20 crc kubenswrapper[4959]: I1007 15:40:20.665937 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-nbdf7_e2e00fa0-bc56-4a3f-9796-bbc1ec5dc47d/controller/0.log" Oct 07 15:40:20 crc kubenswrapper[4959]: I1007 15:40:20.758249 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv4vw_5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f/cp-frr-files/0.log" Oct 07 15:40:20 crc kubenswrapper[4959]: I1007 15:40:20.933979 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv4vw_5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f/cp-reloader/0.log" Oct 07 15:40:20 crc kubenswrapper[4959]: I1007 15:40:20.946185 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv4vw_5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f/cp-metrics/0.log" Oct 07 15:40:20 crc kubenswrapper[4959]: I1007 15:40:20.984060 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv4vw_5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f/cp-frr-files/0.log" Oct 07 15:40:21 crc kubenswrapper[4959]: I1007 15:40:21.059336 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv4vw_5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f/cp-reloader/0.log" Oct 07 15:40:21 crc kubenswrapper[4959]: I1007 15:40:21.141883 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv4vw_5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f/cp-frr-files/0.log" Oct 07 15:40:21 crc kubenswrapper[4959]: I1007 15:40:21.174863 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv4vw_5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f/cp-metrics/0.log" Oct 07 15:40:21 crc kubenswrapper[4959]: I1007 15:40:21.185460 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv4vw_5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f/cp-reloader/0.log" Oct 07 15:40:21 crc kubenswrapper[4959]: I1007 15:40:21.265311 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv4vw_5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f/cp-metrics/0.log" Oct 07 15:40:21 crc kubenswrapper[4959]: I1007 15:40:21.422019 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv4vw_5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f/cp-frr-files/0.log" Oct 07 15:40:21 crc kubenswrapper[4959]: I1007 15:40:21.432871 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv4vw_5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f/cp-metrics/0.log" Oct 07 15:40:21 crc kubenswrapper[4959]: I1007 15:40:21.449616 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv4vw_5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f/cp-reloader/0.log" Oct 07 15:40:21 crc kubenswrapper[4959]: I1007 15:40:21.493513 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv4vw_5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f/controller/0.log" Oct 07 15:40:21 crc kubenswrapper[4959]: I1007 15:40:21.635761 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv4vw_5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f/frr-metrics/0.log" Oct 07 15:40:21 crc kubenswrapper[4959]: I1007 15:40:21.715485 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv4vw_5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f/kube-rbac-proxy-frr/0.log" Oct 07 15:40:21 crc kubenswrapper[4959]: I1007 15:40:21.715670 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv4vw_5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f/kube-rbac-proxy/0.log" Oct 07 15:40:21 crc kubenswrapper[4959]: I1007 15:40:21.917186 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv4vw_5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f/reloader/0.log" Oct 07 15:40:21 crc kubenswrapper[4959]: I1007 15:40:21.944895 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-wcc2g_6dd5ad64-5f6d-4e92-9c1b-3f0837cb134c/frr-k8s-webhook-server/0.log" Oct 07 15:40:22 crc kubenswrapper[4959]: I1007 15:40:22.198308 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-9c45bf8cc-m9pg5_dce9e35b-f4a5-41c7-a47f-f9de5ac4d966/manager/0.log" Oct 07 15:40:22 crc kubenswrapper[4959]: I1007 15:40:22.451975 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-bb2tj_c9ac1753-f8ad-4f22-a955-38663b8121f0/kube-rbac-proxy/0.log" Oct 07 15:40:22 crc kubenswrapper[4959]: I1007 15:40:22.461165 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7d969dc76f-ss6c5_6a48c903-5280-4552-813c-47799ec95dfa/webhook-server/0.log" Oct 07 15:40:23 crc kubenswrapper[4959]: I1007 15:40:23.199619 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-bb2tj_c9ac1753-f8ad-4f22-a955-38663b8121f0/speaker/0.log" Oct 07 15:40:23 crc kubenswrapper[4959]: I1007 15:40:23.742310 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv4vw_5dc42b7f-c0b2-496e-8b60-b7ce5bc5705f/frr/0.log" Oct 07 15:40:35 crc kubenswrapper[4959]: I1007 15:40:35.116214 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c_b524d641-ed92-49cc-b5f1-ff1d8addce42/util/0.log" Oct 07 15:40:35 crc kubenswrapper[4959]: I1007 15:40:35.348755 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c_b524d641-ed92-49cc-b5f1-ff1d8addce42/pull/0.log" Oct 07 15:40:35 crc kubenswrapper[4959]: I1007 15:40:35.391759 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c_b524d641-ed92-49cc-b5f1-ff1d8addce42/pull/0.log" Oct 07 15:40:35 crc kubenswrapper[4959]: I1007 15:40:35.392574 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c_b524d641-ed92-49cc-b5f1-ff1d8addce42/util/0.log" Oct 07 15:40:35 crc kubenswrapper[4959]: I1007 15:40:35.555089 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c_b524d641-ed92-49cc-b5f1-ff1d8addce42/pull/0.log" Oct 07 15:40:35 crc kubenswrapper[4959]: I1007 15:40:35.581720 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c_b524d641-ed92-49cc-b5f1-ff1d8addce42/util/0.log" Oct 07 15:40:35 crc kubenswrapper[4959]: I1007 15:40:35.623040 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2z7j8c_b524d641-ed92-49cc-b5f1-ff1d8addce42/extract/0.log" Oct 07 15:40:35 crc kubenswrapper[4959]: I1007 15:40:35.747556 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dl6rh_0a86d3f1-31eb-4e2e-acdd-592802599d1b/extract-utilities/0.log" Oct 07 15:40:35 crc kubenswrapper[4959]: I1007 15:40:35.957408 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dl6rh_0a86d3f1-31eb-4e2e-acdd-592802599d1b/extract-content/0.log" Oct 07 15:40:35 crc kubenswrapper[4959]: I1007 15:40:35.975754 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dl6rh_0a86d3f1-31eb-4e2e-acdd-592802599d1b/extract-content/0.log" Oct 07 15:40:35 crc kubenswrapper[4959]: I1007 15:40:35.996398 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dl6rh_0a86d3f1-31eb-4e2e-acdd-592802599d1b/extract-utilities/0.log" Oct 07 15:40:36 crc kubenswrapper[4959]: I1007 15:40:36.196538 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dl6rh_0a86d3f1-31eb-4e2e-acdd-592802599d1b/extract-content/0.log" Oct 07 15:40:36 crc kubenswrapper[4959]: I1007 15:40:36.196585 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dl6rh_0a86d3f1-31eb-4e2e-acdd-592802599d1b/extract-utilities/0.log" Oct 07 15:40:36 crc kubenswrapper[4959]: I1007 15:40:36.464075 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-sn6c8_26cb3091-1301-401a-9592-92ba50054762/extract-utilities/0.log" Oct 07 15:40:36 crc kubenswrapper[4959]: I1007 15:40:36.602653 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dl6rh_0a86d3f1-31eb-4e2e-acdd-592802599d1b/registry-server/0.log" Oct 07 15:40:36 crc kubenswrapper[4959]: E1007 15:40:36.654737 4959 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 07 15:40:36 crc kubenswrapper[4959]: I1007 15:40:36.685923 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-sn6c8_26cb3091-1301-401a-9592-92ba50054762/extract-utilities/0.log" Oct 07 15:40:36 crc kubenswrapper[4959]: I1007 15:40:36.692932 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-sn6c8_26cb3091-1301-401a-9592-92ba50054762/extract-content/0.log" Oct 07 15:40:36 crc kubenswrapper[4959]: I1007 15:40:36.762816 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-sn6c8_26cb3091-1301-401a-9592-92ba50054762/extract-content/0.log" Oct 07 15:40:36 crc kubenswrapper[4959]: I1007 15:40:36.953203 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-sn6c8_26cb3091-1301-401a-9592-92ba50054762/extract-content/0.log" Oct 07 15:40:36 crc kubenswrapper[4959]: I1007 15:40:36.966392 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-sn6c8_26cb3091-1301-401a-9592-92ba50054762/extract-utilities/0.log" Oct 07 15:40:37 crc kubenswrapper[4959]: I1007 15:40:37.243555 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz_bbe812c3-c65e-4a52-a87e-7138189e359a/util/0.log" Oct 07 15:40:37 crc kubenswrapper[4959]: I1007 15:40:37.492883 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz_bbe812c3-c65e-4a52-a87e-7138189e359a/pull/0.log" Oct 07 15:40:37 crc kubenswrapper[4959]: I1007 15:40:37.494704 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz_bbe812c3-c65e-4a52-a87e-7138189e359a/pull/0.log" Oct 07 15:40:37 crc kubenswrapper[4959]: I1007 15:40:37.592398 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz_bbe812c3-c65e-4a52-a87e-7138189e359a/util/0.log" Oct 07 15:40:37 crc kubenswrapper[4959]: I1007 15:40:37.823477 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz_bbe812c3-c65e-4a52-a87e-7138189e359a/util/0.log" Oct 07 15:40:37 crc kubenswrapper[4959]: I1007 15:40:37.839673 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz_bbe812c3-c65e-4a52-a87e-7138189e359a/extract/0.log" Oct 07 15:40:37 crc kubenswrapper[4959]: I1007 15:40:37.878826 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7ptcz_bbe812c3-c65e-4a52-a87e-7138189e359a/pull/0.log" Oct 07 15:40:38 crc kubenswrapper[4959]: I1007 15:40:38.107805 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-599kq_922b72b0-51d8-4914-b740-5be4725ab083/marketplace-operator/0.log" Oct 07 15:40:38 crc kubenswrapper[4959]: I1007 15:40:38.221113 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-sn6c8_26cb3091-1301-401a-9592-92ba50054762/registry-server/0.log" Oct 07 15:40:38 crc kubenswrapper[4959]: I1007 15:40:38.350317 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-5tthc_6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c/extract-utilities/0.log" Oct 07 15:40:38 crc kubenswrapper[4959]: I1007 15:40:38.553293 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-5tthc_6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c/extract-content/0.log" Oct 07 15:40:38 crc kubenswrapper[4959]: I1007 15:40:38.559711 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-5tthc_6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c/extract-content/0.log" Oct 07 15:40:38 crc kubenswrapper[4959]: I1007 15:40:38.627188 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-5tthc_6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c/extract-utilities/0.log" Oct 07 15:40:38 crc kubenswrapper[4959]: I1007 15:40:38.814187 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-5tthc_6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c/extract-utilities/0.log" Oct 07 15:40:38 crc kubenswrapper[4959]: I1007 15:40:38.853075 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-5tthc_6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c/extract-content/0.log" Oct 07 15:40:39 crc kubenswrapper[4959]: I1007 15:40:39.024521 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zftzn_e56a45bf-949b-4c96-9978-c8929478ebb7/extract-utilities/0.log" Oct 07 15:40:39 crc kubenswrapper[4959]: I1007 15:40:39.132142 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-5tthc_6fda9f88-41a8-47d3-bdcc-382a0ee9aa2c/registry-server/0.log" Oct 07 15:40:39 crc kubenswrapper[4959]: I1007 15:40:39.251794 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zftzn_e56a45bf-949b-4c96-9978-c8929478ebb7/extract-content/0.log" Oct 07 15:40:39 crc kubenswrapper[4959]: I1007 15:40:39.263219 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zftzn_e56a45bf-949b-4c96-9978-c8929478ebb7/extract-utilities/0.log" Oct 07 15:40:39 crc kubenswrapper[4959]: I1007 15:40:39.265962 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zftzn_e56a45bf-949b-4c96-9978-c8929478ebb7/extract-content/0.log" Oct 07 15:40:39 crc kubenswrapper[4959]: I1007 15:40:39.417106 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zftzn_e56a45bf-949b-4c96-9978-c8929478ebb7/extract-utilities/0.log" Oct 07 15:40:39 crc kubenswrapper[4959]: I1007 15:40:39.419179 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zftzn_e56a45bf-949b-4c96-9978-c8929478ebb7/extract-content/0.log" Oct 07 15:40:39 crc kubenswrapper[4959]: I1007 15:40:39.536709 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zftzn_e56a45bf-949b-4c96-9978-c8929478ebb7/registry-server/0.log" Oct 07 15:40:50 crc kubenswrapper[4959]: I1007 15:40:50.630977 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 15:40:50 crc kubenswrapper[4959]: I1007 15:40:50.632004 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 15:41:20 crc kubenswrapper[4959]: I1007 15:41:20.630879 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 15:41:20 crc kubenswrapper[4959]: I1007 15:41:20.631948 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 15:41:20 crc kubenswrapper[4959]: I1007 15:41:20.632034 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 15:41:20 crc kubenswrapper[4959]: I1007 15:41:20.633419 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6dd605c45ef00573c72aea18d5f4874a4f19110dd7cbf9ba997a985a527c37db"} pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 15:41:20 crc kubenswrapper[4959]: I1007 15:41:20.633493 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" containerID="cri-o://6dd605c45ef00573c72aea18d5f4874a4f19110dd7cbf9ba997a985a527c37db" gracePeriod=600 Oct 07 15:41:21 crc kubenswrapper[4959]: I1007 15:41:21.495199 4959 generic.go:334] "Generic (PLEG): container finished" podID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerID="6dd605c45ef00573c72aea18d5f4874a4f19110dd7cbf9ba997a985a527c37db" exitCode=0 Oct 07 15:41:21 crc kubenswrapper[4959]: I1007 15:41:21.495265 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerDied","Data":"6dd605c45ef00573c72aea18d5f4874a4f19110dd7cbf9ba997a985a527c37db"} Oct 07 15:41:21 crc kubenswrapper[4959]: I1007 15:41:21.496409 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerStarted","Data":"cc4b75248b8dd8f2b094849093a1f96e5ede5f89de9a361482c179cbdf4a1efb"} Oct 07 15:41:21 crc kubenswrapper[4959]: I1007 15:41:21.496458 4959 scope.go:117] "RemoveContainer" containerID="02e62344f2451e337345daacd996959d649b58c3a45eb20cc89ce3b39384958a" Oct 07 15:41:45 crc kubenswrapper[4959]: E1007 15:41:45.654142 4959 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 07 15:43:15 crc kubenswrapper[4959]: E1007 15:43:15.653814 4959 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 07 15:43:19 crc kubenswrapper[4959]: I1007 15:43:19.745822 4959 generic.go:334] "Generic (PLEG): container finished" podID="d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5" containerID="244c8cbefa7361365f17179daec007e4317aa8ef6c39427a8ba5e1719b8b0560" exitCode=0 Oct 07 15:43:19 crc kubenswrapper[4959]: I1007 15:43:19.746029 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tlcvp/must-gather-f4rxm" event={"ID":"d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5","Type":"ContainerDied","Data":"244c8cbefa7361365f17179daec007e4317aa8ef6c39427a8ba5e1719b8b0560"} Oct 07 15:43:19 crc kubenswrapper[4959]: I1007 15:43:19.747031 4959 scope.go:117] "RemoveContainer" containerID="244c8cbefa7361365f17179daec007e4317aa8ef6c39427a8ba5e1719b8b0560" Oct 07 15:43:20 crc kubenswrapper[4959]: I1007 15:43:20.455288 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-tlcvp_must-gather-f4rxm_d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5/gather/0.log" Oct 07 15:43:20 crc kubenswrapper[4959]: I1007 15:43:20.629694 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 15:43:20 crc kubenswrapper[4959]: I1007 15:43:20.629756 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 15:43:29 crc kubenswrapper[4959]: I1007 15:43:29.862118 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-tlcvp/must-gather-f4rxm"] Oct 07 15:43:29 crc kubenswrapper[4959]: I1007 15:43:29.862952 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-tlcvp/must-gather-f4rxm" podUID="d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5" containerName="copy" containerID="cri-o://78053427f49238a92dfe17edbca923b5516edee46ae77d2aa0e841622bc9783b" gracePeriod=2 Oct 07 15:43:29 crc kubenswrapper[4959]: I1007 15:43:29.876911 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-tlcvp/must-gather-f4rxm"] Oct 07 15:43:30 crc kubenswrapper[4959]: I1007 15:43:30.443639 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-tlcvp_must-gather-f4rxm_d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5/copy/0.log" Oct 07 15:43:30 crc kubenswrapper[4959]: I1007 15:43:30.444802 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tlcvp/must-gather-f4rxm" Oct 07 15:43:30 crc kubenswrapper[4959]: I1007 15:43:30.506054 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5-must-gather-output\") pod \"d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5\" (UID: \"d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5\") " Oct 07 15:43:30 crc kubenswrapper[4959]: I1007 15:43:30.506232 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qt7zb\" (UniqueName: \"kubernetes.io/projected/d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5-kube-api-access-qt7zb\") pod \"d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5\" (UID: \"d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5\") " Oct 07 15:43:30 crc kubenswrapper[4959]: I1007 15:43:30.522632 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5-kube-api-access-qt7zb" (OuterVolumeSpecName: "kube-api-access-qt7zb") pod "d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5" (UID: "d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5"). InnerVolumeSpecName "kube-api-access-qt7zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 15:43:30 crc kubenswrapper[4959]: I1007 15:43:30.609470 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qt7zb\" (UniqueName: \"kubernetes.io/projected/d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5-kube-api-access-qt7zb\") on node \"crc\" DevicePath \"\"" Oct 07 15:43:30 crc kubenswrapper[4959]: I1007 15:43:30.721421 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5" (UID: "d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 07 15:43:30 crc kubenswrapper[4959]: I1007 15:43:30.813654 4959 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 07 15:43:30 crc kubenswrapper[4959]: I1007 15:43:30.856805 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-tlcvp_must-gather-f4rxm_d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5/copy/0.log" Oct 07 15:43:30 crc kubenswrapper[4959]: I1007 15:43:30.857354 4959 generic.go:334] "Generic (PLEG): container finished" podID="d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5" containerID="78053427f49238a92dfe17edbca923b5516edee46ae77d2aa0e841622bc9783b" exitCode=143 Oct 07 15:43:30 crc kubenswrapper[4959]: I1007 15:43:30.857485 4959 scope.go:117] "RemoveContainer" containerID="78053427f49238a92dfe17edbca923b5516edee46ae77d2aa0e841622bc9783b" Oct 07 15:43:30 crc kubenswrapper[4959]: I1007 15:43:30.857738 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tlcvp/must-gather-f4rxm" Oct 07 15:43:30 crc kubenswrapper[4959]: I1007 15:43:30.885124 4959 scope.go:117] "RemoveContainer" containerID="244c8cbefa7361365f17179daec007e4317aa8ef6c39427a8ba5e1719b8b0560" Oct 07 15:43:30 crc kubenswrapper[4959]: I1007 15:43:30.945553 4959 scope.go:117] "RemoveContainer" containerID="78053427f49238a92dfe17edbca923b5516edee46ae77d2aa0e841622bc9783b" Oct 07 15:43:30 crc kubenswrapper[4959]: E1007 15:43:30.946333 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78053427f49238a92dfe17edbca923b5516edee46ae77d2aa0e841622bc9783b\": container with ID starting with 78053427f49238a92dfe17edbca923b5516edee46ae77d2aa0e841622bc9783b not found: ID does not exist" containerID="78053427f49238a92dfe17edbca923b5516edee46ae77d2aa0e841622bc9783b" Oct 07 15:43:30 crc kubenswrapper[4959]: I1007 15:43:30.946397 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78053427f49238a92dfe17edbca923b5516edee46ae77d2aa0e841622bc9783b"} err="failed to get container status \"78053427f49238a92dfe17edbca923b5516edee46ae77d2aa0e841622bc9783b\": rpc error: code = NotFound desc = could not find container \"78053427f49238a92dfe17edbca923b5516edee46ae77d2aa0e841622bc9783b\": container with ID starting with 78053427f49238a92dfe17edbca923b5516edee46ae77d2aa0e841622bc9783b not found: ID does not exist" Oct 07 15:43:30 crc kubenswrapper[4959]: I1007 15:43:30.946443 4959 scope.go:117] "RemoveContainer" containerID="244c8cbefa7361365f17179daec007e4317aa8ef6c39427a8ba5e1719b8b0560" Oct 07 15:43:30 crc kubenswrapper[4959]: E1007 15:43:30.947022 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"244c8cbefa7361365f17179daec007e4317aa8ef6c39427a8ba5e1719b8b0560\": container with ID starting with 244c8cbefa7361365f17179daec007e4317aa8ef6c39427a8ba5e1719b8b0560 not found: ID does not exist" containerID="244c8cbefa7361365f17179daec007e4317aa8ef6c39427a8ba5e1719b8b0560" Oct 07 15:43:30 crc kubenswrapper[4959]: I1007 15:43:30.947056 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"244c8cbefa7361365f17179daec007e4317aa8ef6c39427a8ba5e1719b8b0560"} err="failed to get container status \"244c8cbefa7361365f17179daec007e4317aa8ef6c39427a8ba5e1719b8b0560\": rpc error: code = NotFound desc = could not find container \"244c8cbefa7361365f17179daec007e4317aa8ef6c39427a8ba5e1719b8b0560\": container with ID starting with 244c8cbefa7361365f17179daec007e4317aa8ef6c39427a8ba5e1719b8b0560 not found: ID does not exist" Oct 07 15:43:32 crc kubenswrapper[4959]: I1007 15:43:32.667057 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5" path="/var/lib/kubelet/pods/d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5/volumes" Oct 07 15:43:50 crc kubenswrapper[4959]: I1007 15:43:50.630076 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 15:43:50 crc kubenswrapper[4959]: I1007 15:43:50.631187 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 15:44:20 crc kubenswrapper[4959]: I1007 15:44:20.630571 4959 patch_prober.go:28] interesting pod/machine-config-daemon-m5zzg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 07 15:44:20 crc kubenswrapper[4959]: I1007 15:44:20.631359 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 07 15:44:20 crc kubenswrapper[4959]: I1007 15:44:20.631425 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" Oct 07 15:44:20 crc kubenswrapper[4959]: I1007 15:44:20.632217 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cc4b75248b8dd8f2b094849093a1f96e5ede5f89de9a361482c179cbdf4a1efb"} pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 07 15:44:20 crc kubenswrapper[4959]: I1007 15:44:20.632266 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerName="machine-config-daemon" containerID="cri-o://cc4b75248b8dd8f2b094849093a1f96e5ede5f89de9a361482c179cbdf4a1efb" gracePeriod=600 Oct 07 15:44:20 crc kubenswrapper[4959]: E1007 15:44:20.774020 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:44:21 crc kubenswrapper[4959]: I1007 15:44:21.383698 4959 generic.go:334] "Generic (PLEG): container finished" podID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" containerID="cc4b75248b8dd8f2b094849093a1f96e5ede5f89de9a361482c179cbdf4a1efb" exitCode=0 Oct 07 15:44:21 crc kubenswrapper[4959]: I1007 15:44:21.383822 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" event={"ID":"0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a","Type":"ContainerDied","Data":"cc4b75248b8dd8f2b094849093a1f96e5ede5f89de9a361482c179cbdf4a1efb"} Oct 07 15:44:21 crc kubenswrapper[4959]: I1007 15:44:21.384387 4959 scope.go:117] "RemoveContainer" containerID="6dd605c45ef00573c72aea18d5f4874a4f19110dd7cbf9ba997a985a527c37db" Oct 07 15:44:21 crc kubenswrapper[4959]: I1007 15:44:21.386115 4959 scope.go:117] "RemoveContainer" containerID="cc4b75248b8dd8f2b094849093a1f96e5ede5f89de9a361482c179cbdf4a1efb" Oct 07 15:44:21 crc kubenswrapper[4959]: E1007 15:44:21.386968 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:44:34 crc kubenswrapper[4959]: E1007 15:44:34.653722 4959 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 07 15:44:34 crc kubenswrapper[4959]: I1007 15:44:34.654000 4959 scope.go:117] "RemoveContainer" containerID="cc4b75248b8dd8f2b094849093a1f96e5ede5f89de9a361482c179cbdf4a1efb" Oct 07 15:44:34 crc kubenswrapper[4959]: E1007 15:44:34.654730 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:44:46 crc kubenswrapper[4959]: I1007 15:44:46.654030 4959 scope.go:117] "RemoveContainer" containerID="cc4b75248b8dd8f2b094849093a1f96e5ede5f89de9a361482c179cbdf4a1efb" Oct 07 15:44:46 crc kubenswrapper[4959]: E1007 15:44:46.655146 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:44:57 crc kubenswrapper[4959]: I1007 15:44:57.653985 4959 scope.go:117] "RemoveContainer" containerID="cc4b75248b8dd8f2b094849093a1f96e5ede5f89de9a361482c179cbdf4a1efb" Oct 07 15:44:57 crc kubenswrapper[4959]: E1007 15:44:57.655452 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:45:00 crc kubenswrapper[4959]: I1007 15:45:00.167493 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330865-479vd"] Oct 07 15:45:00 crc kubenswrapper[4959]: E1007 15:45:00.168547 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5" containerName="gather" Oct 07 15:45:00 crc kubenswrapper[4959]: I1007 15:45:00.168565 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5" containerName="gather" Oct 07 15:45:00 crc kubenswrapper[4959]: E1007 15:45:00.168600 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dabfbd4d-255f-45f7-ab12-6450a2d723b2" containerName="container-00" Oct 07 15:45:00 crc kubenswrapper[4959]: I1007 15:45:00.168606 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="dabfbd4d-255f-45f7-ab12-6450a2d723b2" containerName="container-00" Oct 07 15:45:00 crc kubenswrapper[4959]: E1007 15:45:00.168634 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5" containerName="copy" Oct 07 15:45:00 crc kubenswrapper[4959]: I1007 15:45:00.168640 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5" containerName="copy" Oct 07 15:45:00 crc kubenswrapper[4959]: I1007 15:45:00.168894 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5" containerName="gather" Oct 07 15:45:00 crc kubenswrapper[4959]: I1007 15:45:00.168915 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7eb7d0a-cf24-464a-a446-6b6ccd53ccd5" containerName="copy" Oct 07 15:45:00 crc kubenswrapper[4959]: I1007 15:45:00.168942 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="dabfbd4d-255f-45f7-ab12-6450a2d723b2" containerName="container-00" Oct 07 15:45:00 crc kubenswrapper[4959]: I1007 15:45:00.169770 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330865-479vd" Oct 07 15:45:00 crc kubenswrapper[4959]: I1007 15:45:00.172995 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 07 15:45:00 crc kubenswrapper[4959]: I1007 15:45:00.175202 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 07 15:45:00 crc kubenswrapper[4959]: I1007 15:45:00.179955 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330865-479vd"] Oct 07 15:45:00 crc kubenswrapper[4959]: I1007 15:45:00.286301 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5302adf-15d8-4809-bf0c-95ff460ce841-secret-volume\") pod \"collect-profiles-29330865-479vd\" (UID: \"a5302adf-15d8-4809-bf0c-95ff460ce841\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330865-479vd" Oct 07 15:45:00 crc kubenswrapper[4959]: I1007 15:45:00.286454 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4xrb\" (UniqueName: \"kubernetes.io/projected/a5302adf-15d8-4809-bf0c-95ff460ce841-kube-api-access-q4xrb\") pod \"collect-profiles-29330865-479vd\" (UID: \"a5302adf-15d8-4809-bf0c-95ff460ce841\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330865-479vd" Oct 07 15:45:00 crc kubenswrapper[4959]: I1007 15:45:00.286498 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5302adf-15d8-4809-bf0c-95ff460ce841-config-volume\") pod \"collect-profiles-29330865-479vd\" (UID: \"a5302adf-15d8-4809-bf0c-95ff460ce841\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330865-479vd" Oct 07 15:45:00 crc kubenswrapper[4959]: I1007 15:45:00.389603 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5302adf-15d8-4809-bf0c-95ff460ce841-config-volume\") pod \"collect-profiles-29330865-479vd\" (UID: \"a5302adf-15d8-4809-bf0c-95ff460ce841\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330865-479vd" Oct 07 15:45:00 crc kubenswrapper[4959]: I1007 15:45:00.390231 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5302adf-15d8-4809-bf0c-95ff460ce841-secret-volume\") pod \"collect-profiles-29330865-479vd\" (UID: \"a5302adf-15d8-4809-bf0c-95ff460ce841\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330865-479vd" Oct 07 15:45:00 crc kubenswrapper[4959]: I1007 15:45:00.390418 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4xrb\" (UniqueName: \"kubernetes.io/projected/a5302adf-15d8-4809-bf0c-95ff460ce841-kube-api-access-q4xrb\") pod \"collect-profiles-29330865-479vd\" (UID: \"a5302adf-15d8-4809-bf0c-95ff460ce841\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330865-479vd" Oct 07 15:45:00 crc kubenswrapper[4959]: I1007 15:45:00.390743 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5302adf-15d8-4809-bf0c-95ff460ce841-config-volume\") pod \"collect-profiles-29330865-479vd\" (UID: \"a5302adf-15d8-4809-bf0c-95ff460ce841\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330865-479vd" Oct 07 15:45:00 crc kubenswrapper[4959]: I1007 15:45:00.400503 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5302adf-15d8-4809-bf0c-95ff460ce841-secret-volume\") pod \"collect-profiles-29330865-479vd\" (UID: \"a5302adf-15d8-4809-bf0c-95ff460ce841\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330865-479vd" Oct 07 15:45:00 crc kubenswrapper[4959]: I1007 15:45:00.421883 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4xrb\" (UniqueName: \"kubernetes.io/projected/a5302adf-15d8-4809-bf0c-95ff460ce841-kube-api-access-q4xrb\") pod \"collect-profiles-29330865-479vd\" (UID: \"a5302adf-15d8-4809-bf0c-95ff460ce841\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29330865-479vd" Oct 07 15:45:00 crc kubenswrapper[4959]: I1007 15:45:00.495300 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330865-479vd" Oct 07 15:45:00 crc kubenswrapper[4959]: I1007 15:45:00.982384 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330865-479vd"] Oct 07 15:45:01 crc kubenswrapper[4959]: I1007 15:45:01.781985 4959 generic.go:334] "Generic (PLEG): container finished" podID="a5302adf-15d8-4809-bf0c-95ff460ce841" containerID="63ca3e87d056ac136be8af80dfba029c4cde52757822585b755eb35fa6647c9d" exitCode=0 Oct 07 15:45:01 crc kubenswrapper[4959]: I1007 15:45:01.782038 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330865-479vd" event={"ID":"a5302adf-15d8-4809-bf0c-95ff460ce841","Type":"ContainerDied","Data":"63ca3e87d056ac136be8af80dfba029c4cde52757822585b755eb35fa6647c9d"} Oct 07 15:45:01 crc kubenswrapper[4959]: I1007 15:45:01.782069 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330865-479vd" event={"ID":"a5302adf-15d8-4809-bf0c-95ff460ce841","Type":"ContainerStarted","Data":"cbc79c915d11ccdd599f3dcbcf5d44f115208f23c8e1ef6fe83bd58fcade810c"} Oct 07 15:45:03 crc kubenswrapper[4959]: I1007 15:45:03.141709 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330865-479vd" Oct 07 15:45:03 crc kubenswrapper[4959]: I1007 15:45:03.252410 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5302adf-15d8-4809-bf0c-95ff460ce841-secret-volume\") pod \"a5302adf-15d8-4809-bf0c-95ff460ce841\" (UID: \"a5302adf-15d8-4809-bf0c-95ff460ce841\") " Oct 07 15:45:03 crc kubenswrapper[4959]: I1007 15:45:03.252544 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4xrb\" (UniqueName: \"kubernetes.io/projected/a5302adf-15d8-4809-bf0c-95ff460ce841-kube-api-access-q4xrb\") pod \"a5302adf-15d8-4809-bf0c-95ff460ce841\" (UID: \"a5302adf-15d8-4809-bf0c-95ff460ce841\") " Oct 07 15:45:03 crc kubenswrapper[4959]: I1007 15:45:03.252764 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5302adf-15d8-4809-bf0c-95ff460ce841-config-volume\") pod \"a5302adf-15d8-4809-bf0c-95ff460ce841\" (UID: \"a5302adf-15d8-4809-bf0c-95ff460ce841\") " Oct 07 15:45:03 crc kubenswrapper[4959]: I1007 15:45:03.253754 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5302adf-15d8-4809-bf0c-95ff460ce841-config-volume" (OuterVolumeSpecName: "config-volume") pod "a5302adf-15d8-4809-bf0c-95ff460ce841" (UID: "a5302adf-15d8-4809-bf0c-95ff460ce841"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 07 15:45:03 crc kubenswrapper[4959]: I1007 15:45:03.258074 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5302adf-15d8-4809-bf0c-95ff460ce841-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a5302adf-15d8-4809-bf0c-95ff460ce841" (UID: "a5302adf-15d8-4809-bf0c-95ff460ce841"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 07 15:45:03 crc kubenswrapper[4959]: I1007 15:45:03.266093 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5302adf-15d8-4809-bf0c-95ff460ce841-kube-api-access-q4xrb" (OuterVolumeSpecName: "kube-api-access-q4xrb") pod "a5302adf-15d8-4809-bf0c-95ff460ce841" (UID: "a5302adf-15d8-4809-bf0c-95ff460ce841"). InnerVolumeSpecName "kube-api-access-q4xrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 07 15:45:03 crc kubenswrapper[4959]: I1007 15:45:03.355683 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4xrb\" (UniqueName: \"kubernetes.io/projected/a5302adf-15d8-4809-bf0c-95ff460ce841-kube-api-access-q4xrb\") on node \"crc\" DevicePath \"\"" Oct 07 15:45:03 crc kubenswrapper[4959]: I1007 15:45:03.356306 4959 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5302adf-15d8-4809-bf0c-95ff460ce841-config-volume\") on node \"crc\" DevicePath \"\"" Oct 07 15:45:03 crc kubenswrapper[4959]: I1007 15:45:03.356401 4959 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5302adf-15d8-4809-bf0c-95ff460ce841-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 07 15:45:03 crc kubenswrapper[4959]: I1007 15:45:03.806860 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29330865-479vd" event={"ID":"a5302adf-15d8-4809-bf0c-95ff460ce841","Type":"ContainerDied","Data":"cbc79c915d11ccdd599f3dcbcf5d44f115208f23c8e1ef6fe83bd58fcade810c"} Oct 07 15:45:03 crc kubenswrapper[4959]: I1007 15:45:03.806912 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbc79c915d11ccdd599f3dcbcf5d44f115208f23c8e1ef6fe83bd58fcade810c" Oct 07 15:45:03 crc kubenswrapper[4959]: I1007 15:45:03.806978 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29330865-479vd" Oct 07 15:45:04 crc kubenswrapper[4959]: E1007 15:45:04.022142 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5302adf_15d8_4809_bf0c_95ff460ce841.slice/crio-cbc79c915d11ccdd599f3dcbcf5d44f115208f23c8e1ef6fe83bd58fcade810c\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5302adf_15d8_4809_bf0c_95ff460ce841.slice\": RecentStats: unable to find data in memory cache]" Oct 07 15:45:04 crc kubenswrapper[4959]: I1007 15:45:04.221950 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330820-j5lrc"] Oct 07 15:45:04 crc kubenswrapper[4959]: I1007 15:45:04.231999 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29330820-j5lrc"] Oct 07 15:45:04 crc kubenswrapper[4959]: I1007 15:45:04.667249 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63f04186-4fb9-4518-b19e-b0a7bf573df0" path="/var/lib/kubelet/pods/63f04186-4fb9-4518-b19e-b0a7bf573df0/volumes" Oct 07 15:45:12 crc kubenswrapper[4959]: I1007 15:45:12.654788 4959 scope.go:117] "RemoveContainer" containerID="cc4b75248b8dd8f2b094849093a1f96e5ede5f89de9a361482c179cbdf4a1efb" Oct 07 15:45:12 crc kubenswrapper[4959]: E1007 15:45:12.655463 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:45:26 crc kubenswrapper[4959]: I1007 15:45:26.657233 4959 scope.go:117] "RemoveContainer" containerID="cc4b75248b8dd8f2b094849093a1f96e5ede5f89de9a361482c179cbdf4a1efb" Oct 07 15:45:26 crc kubenswrapper[4959]: E1007 15:45:26.658378 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:45:37 crc kubenswrapper[4959]: I1007 15:45:37.194452 4959 scope.go:117] "RemoveContainer" containerID="d9ff643ce7ebe5d9592f4e04104a93b0cc3a3afc9298f8db3a8eedd592a8330d" Oct 07 15:45:37 crc kubenswrapper[4959]: I1007 15:45:37.219030 4959 scope.go:117] "RemoveContainer" containerID="049810a597be74998d30a03070d4f11a3c8ae4ec242bc4ede29616ab3dfc947a" Oct 07 15:45:40 crc kubenswrapper[4959]: I1007 15:45:40.653853 4959 scope.go:117] "RemoveContainer" containerID="cc4b75248b8dd8f2b094849093a1f96e5ede5f89de9a361482c179cbdf4a1efb" Oct 07 15:45:40 crc kubenswrapper[4959]: E1007 15:45:40.654436 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:45:42 crc kubenswrapper[4959]: E1007 15:45:42.653735 4959 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 07 15:45:55 crc kubenswrapper[4959]: I1007 15:45:55.654521 4959 scope.go:117] "RemoveContainer" containerID="cc4b75248b8dd8f2b094849093a1f96e5ede5f89de9a361482c179cbdf4a1efb" Oct 07 15:45:55 crc kubenswrapper[4959]: E1007 15:45:55.655690 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" Oct 07 15:46:07 crc kubenswrapper[4959]: I1007 15:46:07.653275 4959 scope.go:117] "RemoveContainer" containerID="cc4b75248b8dd8f2b094849093a1f96e5ede5f89de9a361482c179cbdf4a1efb" Oct 07 15:46:07 crc kubenswrapper[4959]: E1007 15:46:07.655353 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-m5zzg_openshift-machine-config-operator(0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a)\"" pod="openshift-machine-config-operator/machine-config-daemon-m5zzg" podUID="0ac172b2-9dfc-446a-ba8f-6f1cb4df4c4a" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515071232720024444 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015071232721017362 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015071214216016504 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015071214216015454 5ustar corecore